# include "config.h"
#endif
+#include <string.h>
+
#include "video.h"
+static int get_size (GstVideoFormat format, int width, int height);
+static int get_stride (GstVideoFormat format, int plane, int width);
+static int fill_planes (GstVideoInfo * info);
+
+typedef enum
+{
+ VIDEO_FLAG_YUV = (1 << 0),
+ VIDEO_FLAG_RGB = (1 << 1),
+ VIDEO_FLAG_GRAY = (1 << 2),
+ VIDEO_FLAG_ALPHA = (1 << 3)
+} VideoFlags;
+
+typedef struct
+{
+ const gchar *fmt;
+ GstVideoFormat format;
+ guint32 fourcc;
+ VideoFlags flags;
+ guint n_comp;
+ guint depth[GST_VIDEO_MAX_PLANES];
+ guint n_planes;
+} VideoFormat;
+
+#define COMP0 0, { 0, 0, 0, 0 }
+#define COMP8 1, { 8, 0, 0, 0 }
+#define COMP888 3, { 8, 8, 8, 0 }
+#define COMP8888 4, { 8, 8, 8, 8 }
+#define COMP10_10_10 3, { 10, 10, 10, 0 }
+#define COMP16 1, { 16, 0, 0, 0 }
+#define COMP16_16_16 3, { 16, 16, 16, 0 }
+#define COMP16_16_16_16 4, { 16, 16, 16, 16 }
+#define COMP555 3, { 5, 5, 5, 0 }
+#define COMP565 3, { 5, 6, 5, 0 }
+
+#define MAKE_YUV_FORMAT(name, fourcc, comp) \
+ { G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, fourcc, VIDEO_FLAG_YUV, comp }
+#define MAKE_YUVA_FORMAT(name, fourcc, comp) \
+ { G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, fourcc, VIDEO_FLAG_YUV | VIDEO_FLAG_ALPHA, comp }
+
+#define MAKE_RGB_FORMAT(name, comp) \
+ { G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, 0x00000000, VIDEO_FLAG_RGB, comp }
+#define MAKE_RGBA_FORMAT(name, comp) \
+ { G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, 0x00000000, VIDEO_FLAG_RGB | VIDEO_FLAG_ALPHA, comp }
+
+#define MAKE_GRAY_FORMAT(name, comp) \
+ { G_STRINGIFY(name), GST_VIDEO_FORMAT_ ##name, 0x00000000, VIDEO_FLAG_GRAY, comp }
+
+static VideoFormat formats[] = {
+ {"UNKNOWN", GST_VIDEO_FORMAT_UNKNOWN, 0x00000000, 0, COMP0},
+
+ MAKE_YUV_FORMAT (I420, GST_MAKE_FOURCC ('I', '4', '2', '0'), COMP888),
+ MAKE_YUV_FORMAT (YV12, GST_MAKE_FOURCC ('Y', 'V', '1', '2'), COMP888),
+ MAKE_YUV_FORMAT (YUY2, GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'), COMP888),
+ MAKE_YUV_FORMAT (UYVY, GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'), COMP888),
+ MAKE_YUVA_FORMAT (AYUV, GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'), COMP8888),
+ MAKE_RGB_FORMAT (RGBx, COMP888),
+ MAKE_RGB_FORMAT (BGRx, COMP888),
+ MAKE_RGB_FORMAT (xRGB, COMP888),
+ MAKE_RGB_FORMAT (xBGR, COMP888),
+ MAKE_RGBA_FORMAT (RGBA, COMP8888),
+ MAKE_RGBA_FORMAT (BGRA, COMP8888),
+ MAKE_RGBA_FORMAT (ARGB, COMP8888),
+ MAKE_RGBA_FORMAT (ABGR, COMP8888),
+ MAKE_RGB_FORMAT (RGB, COMP888),
+ MAKE_RGB_FORMAT (BGR, COMP888),
+
+ MAKE_YUV_FORMAT (Y41B, GST_MAKE_FOURCC ('Y', '4', '1', 'B'), COMP888),
+ MAKE_YUV_FORMAT (Y42B, GST_MAKE_FOURCC ('Y', '4', '2', 'B'), COMP888),
+ MAKE_YUV_FORMAT (YVYU, GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'), COMP888),
+ MAKE_YUV_FORMAT (Y444, GST_MAKE_FOURCC ('Y', '4', '4', '4'), COMP888),
+ MAKE_YUV_FORMAT (v210, GST_MAKE_FOURCC ('v', '2', '1', '0'), COMP10_10_10),
+ MAKE_YUV_FORMAT (v216, GST_MAKE_FOURCC ('v', '2', '1', '6'), COMP16_16_16),
+ MAKE_YUV_FORMAT (NV12, GST_MAKE_FOURCC ('N', 'V', '1', '2'), COMP888),
+ MAKE_YUV_FORMAT (NV21, GST_MAKE_FOURCC ('N', 'V', '2', '1'), COMP888),
+
+ MAKE_GRAY_FORMAT (GRAY8, COMP8),
+ MAKE_GRAY_FORMAT (GRAY16_BE, COMP16),
+ MAKE_GRAY_FORMAT (GRAY16_LE, COMP16),
+
+ MAKE_YUV_FORMAT (v308, GST_MAKE_FOURCC ('v', '3', '0', '8'), COMP888),
+ MAKE_YUV_FORMAT (Y800, GST_MAKE_FOURCC ('Y', '8', '0', '0'), COMP8),
+ MAKE_YUV_FORMAT (Y16, GST_MAKE_FOURCC ('Y', '1', '6', ' '), COMP16),
+
+ MAKE_RGB_FORMAT (RGB16, COMP565),
+ MAKE_RGB_FORMAT (BGR16, COMP565),
+ MAKE_RGB_FORMAT (RGB15, COMP555),
+ MAKE_RGB_FORMAT (BGR15, COMP555),
+
+ MAKE_YUV_FORMAT (UYVP, GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'), COMP10_10_10),
+ MAKE_YUVA_FORMAT (A420, GST_MAKE_FOURCC ('A', '4', '2', '0'), COMP888),
+ MAKE_RGBA_FORMAT (RGB8_PALETTED, COMP8888),
+ MAKE_YUV_FORMAT (YUV9, GST_MAKE_FOURCC ('Y', 'U', 'V', '9'), COMP888),
+ MAKE_YUV_FORMAT (YVU9, GST_MAKE_FOURCC ('Y', 'V', 'U', '9'), COMP888),
+ MAKE_YUV_FORMAT (IYU1, GST_MAKE_FOURCC ('I', 'Y', 'U', '1'), COMP888),
+ MAKE_RGBA_FORMAT (ARGB64, COMP16_16_16_16),
+ MAKE_YUVA_FORMAT (AYUV64, 0x00000000, COMP16_16_16_16),
+ MAKE_YUV_FORMAT (r210, GST_MAKE_FOURCC ('r', '2', '1', '0'), COMP10_10_10),
+};
+
/**
* SECTION:gstvideo
* @short_description: Support library for video operations
*
* <refsect2>
* <para>
- * This library contains some helper functions and includes the
+ * This library contains some helper functions and includes the
* videosink and videofilter base classes.
* </para>
* </refsect2>
*/
-static GstVideoFormat gst_video_format_from_rgb32_masks (int red_mask,
- int green_mask, int blue_mask);
-static GstVideoFormat gst_video_format_from_rgba32_masks (int red_mask,
- int green_mask, int blue_mask, int alpha_mask);
-static GstVideoFormat gst_video_format_from_rgb24_masks (int red_mask,
- int green_mask, int blue_mask);
-
-
-/**
- * gst_video_frame_rate:
- * @pad: pointer to a #GstPad
- *
- * A convenience function to retrieve a GValue holding the framerate
- * from the caps on a pad.
- *
- * The pad needs to have negotiated caps containing a framerate property.
- *
- * Returns: NULL if the pad has no configured caps or the configured caps
- * do not contain a framerate.
- *
- */
-const GValue *
-gst_video_frame_rate (GstPad * pad)
-{
- const GValue *fps;
- gchar *fps_string;
-
- const GstCaps *caps = NULL;
- GstStructure *structure;
-
- /* get pad caps */
- caps = GST_PAD_CAPS (pad);
- if (caps == NULL) {
- g_warning ("gstvideo: failed to get caps of pad %s:%s",
- GST_DEBUG_PAD_NAME (pad));
- return NULL;
- }
-
- structure = gst_caps_get_structure (caps, 0);
- if ((fps = gst_structure_get_value (structure, "framerate")) == NULL) {
- g_warning ("gstvideo: failed to get framerate property of pad %s:%s",
- GST_DEBUG_PAD_NAME (pad));
- return NULL;
- }
- if (!GST_VALUE_HOLDS_FRACTION (fps)) {
- g_warning
- ("gstvideo: framerate property of pad %s:%s is not of type Fraction",
- GST_DEBUG_PAD_NAME (pad));
- return NULL;
- }
-
- fps_string = gst_value_serialize (fps);
- GST_DEBUG ("Framerate request on pad %s:%s: %s",
- GST_DEBUG_PAD_NAME (pad), fps_string);
- g_free (fps_string);
-
- return fps;
-}
-
-/**
- * gst_video_get_size:
- * @pad: pointer to a #GstPad
- * @width: pointer to integer to hold pixel width of the video frames (output)
- * @height: pointer to integer to hold pixel height of the video frames (output)
- *
- * Inspect the caps of the provided pad and retrieve the width and height of
- * the video frames it is configured for.
- *
- * The pad needs to have negotiated caps containing width and height properties.
- *
- * Returns: TRUE if the width and height could be retrieved.
- *
- */
-gboolean
-gst_video_get_size (GstPad * pad, gint * width, gint * height)
-{
- const GstCaps *caps = NULL;
- GstStructure *structure;
- gboolean ret;
-
- g_return_val_if_fail (pad != NULL, FALSE);
- g_return_val_if_fail (width != NULL, FALSE);
- g_return_val_if_fail (height != NULL, FALSE);
-
- caps = GST_PAD_CAPS (pad);
-
- if (caps == NULL) {
- g_warning ("gstvideo: failed to get caps of pad %s:%s",
- GST_DEBUG_PAD_NAME (pad));
- return FALSE;
- }
-
- structure = gst_caps_get_structure (caps, 0);
- ret = gst_structure_get_int (structure, "width", width);
- ret &= gst_structure_get_int (structure, "height", height);
-
- if (!ret) {
- g_warning ("gstvideo: failed to get size properties on pad %s:%s",
- GST_DEBUG_PAD_NAME (pad));
- return FALSE;
- }
-
- GST_DEBUG ("size request on pad %s:%s: %dx%d",
- GST_DEBUG_PAD_NAME (pad), width ? *width : -1, height ? *height : -1);
-
- return TRUE;
-}
-
/**
* gst_video_calculate_display_ratio:
* @dar_n: Numerator of the calculated display_ratio
* @display_par_n: Numerator of the pixel aspect ratio of the display device
* @display_par_d: Denominator of the pixel aspect ratio of the display device
*
- * Given the Pixel Aspect Ratio and size of an input video frame, and the
- * pixel aspect ratio of the intended display device, calculates the actual
+ * Given the Pixel Aspect Ratio and size of an input video frame, and the
+ * pixel aspect ratio of the intended display device, calculates the actual
* display ratio the video will be rendered with.
*
- * Returns: A boolean indicating success and a calculated Display Ratio in the
- * dar_n and dar_d parameters.
- * The return value is FALSE in the case of integer overflow or other error.
+ * Returns: A boolean indicating success and a calculated Display Ratio in the
+ * dar_n and dar_d parameters.
+ * The return value is FALSE in the case of integer overflow or other error.
*
* Since: 0.10.7
*/
guint display_par_n, guint display_par_d)
{
gint num, den;
-
- GValue display_ratio = { 0, };
- GValue tmp = { 0, };
- GValue tmp2 = { 0, };
+ gint tmp_n, tmp_d;
g_return_val_if_fail (dar_n != NULL, FALSE);
g_return_val_if_fail (dar_d != NULL, FALSE);
- g_value_init (&display_ratio, GST_TYPE_FRACTION);
- g_value_init (&tmp, GST_TYPE_FRACTION);
- g_value_init (&tmp2, GST_TYPE_FRACTION);
-
/* Calculate (video_width * video_par_n * display_par_d) /
* (video_height * video_par_d * display_par_n) */
- gst_value_set_fraction (&display_ratio, video_width, video_height);
- gst_value_set_fraction (&tmp, video_par_n, video_par_d);
-
- if (!gst_value_fraction_multiply (&tmp2, &display_ratio, &tmp))
+ if (!gst_util_fraction_multiply (video_width, video_height, video_par_n,
+ video_par_d, &tmp_n, &tmp_d))
goto error_overflow;
- gst_value_set_fraction (&tmp, display_par_d, display_par_n);
-
- if (!gst_value_fraction_multiply (&display_ratio, &tmp2, &tmp))
+ if (!gst_util_fraction_multiply (tmp_n, tmp_d, display_par_d, display_par_n,
+ &num, &den))
goto error_overflow;
- num = gst_value_get_fraction_numerator (&display_ratio);
- den = gst_value_get_fraction_denominator (&display_ratio);
-
- g_value_unset (&display_ratio);
- g_value_unset (&tmp);
- g_value_unset (&tmp2);
-
g_return_val_if_fail (num > 0, FALSE);
g_return_val_if_fail (den > 0, FALSE);
*dar_d = den;
return TRUE;
-error_overflow:
- g_value_unset (&display_ratio);
- g_value_unset (&tmp);
- g_value_unset (&tmp2);
- return FALSE;
-}
-/**
- * gst_video_format_parse_caps_interlaced:
- * @caps: the fixed #GstCaps to parse
- * @interlaced: whether @caps represents interlaced video or not, may be NULL (output)
- *
- * Extracts whether the caps represents interlaced content or not and places it
- * in @interlaced.
- *
- * Since: 0.10.23
- *
- * Returns: TRUE if @caps was parsed correctly.
- */
-gboolean
-gst_video_format_parse_caps_interlaced (GstCaps * caps, gboolean * interlaced)
-{
- GstStructure *structure;
-
- if (!gst_caps_is_fixed (caps))
+ /* ERRORS */
+error_overflow:
+ {
+ GST_WARNING ("overflow in multiply");
return FALSE;
-
- structure = gst_caps_get_structure (caps, 0);
-
- if (interlaced) {
- if (!gst_structure_get_boolean (structure, "interlaced", interlaced))
- *interlaced = FALSE;
}
-
- return TRUE;
}
-/**
- * gst_video_format_parse_caps:
- * @caps: the #GstCaps to parse
- * @format: the #GstVideoFormat of the video represented by @caps (output)
- * @width: the width of the video represented by @caps, may be NULL (output)
- * @height: the height of the video represented by @caps, may be NULL (output)
- *
- * Determines the #GstVideoFormat of @caps and places it in the location
- * pointed to by @format. Extracts the size of the video and places it
- * in the location pointed to by @width and @height. If @caps does not
- * represent one of the raw video formats listed in #GstVideoFormat, the
- * function will fail and return FALSE.
- *
- * Since: 0.10.16
- *
- * Returns: TRUE if @caps was parsed correctly.
- */
-gboolean
-gst_video_format_parse_caps (GstCaps * caps, GstVideoFormat * format,
- int *width, int *height)
+static GstVideoFormat
+gst_video_format_from_rgb32_masks (int red_mask, int green_mask, int blue_mask)
{
- GstStructure *structure;
- gboolean ok = TRUE;
-
- if (!gst_caps_is_fixed (caps))
- return FALSE;
-
- structure = gst_caps_get_structure (caps, 0);
-
- if (format) {
- if (gst_structure_has_name (structure, "video/x-raw-yuv")) {
- guint32 fourcc;
-
- ok &= gst_structure_get_fourcc (structure, "format", &fourcc);
-
- *format = gst_video_format_from_fourcc (fourcc);
- if (*format == GST_VIDEO_FORMAT_UNKNOWN) {
- ok = FALSE;
- }
- } else if (gst_structure_has_name (structure, "video/x-raw-rgb")) {
- int depth;
- int bpp;
- int endianness;
- int red_mask;
- int green_mask;
- int blue_mask;
- int alpha_mask;
- gboolean have_alpha;
-
- ok &= gst_structure_get_int (structure, "depth", &depth);
- ok &= gst_structure_get_int (structure, "bpp", &bpp);
- ok &= gst_structure_get_int (structure, "endianness", &endianness);
- ok &= gst_structure_get_int (structure, "red_mask", &red_mask);
- ok &= gst_structure_get_int (structure, "green_mask", &green_mask);
- ok &= gst_structure_get_int (structure, "blue_mask", &blue_mask);
- have_alpha = gst_structure_get_int (structure, "alpha_mask", &alpha_mask);
-
- if (depth == 24 && bpp == 32 && endianness == G_BIG_ENDIAN) {
- *format = gst_video_format_from_rgb32_masks (red_mask, green_mask,
- blue_mask);
- if (*format == GST_VIDEO_FORMAT_UNKNOWN) {
- ok = FALSE;
- }
- } else if (depth == 32 && bpp == 32 && endianness == G_BIG_ENDIAN &&
- have_alpha) {
- *format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
- blue_mask, alpha_mask);
- if (*format == GST_VIDEO_FORMAT_UNKNOWN) {
- ok = FALSE;
- }
- } else if (depth == 24 && bpp == 24 && endianness == G_BIG_ENDIAN) {
- *format = gst_video_format_from_rgb24_masks (red_mask, green_mask,
- blue_mask);
- if (*format == GST_VIDEO_FORMAT_UNKNOWN) {
- ok = FALSE;
- }
- } else {
- ok = FALSE;
- }
- } else {
- ok = FALSE;
- }
+ if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
+ blue_mask == 0x0000ff00) {
+ return GST_VIDEO_FORMAT_RGBx;
}
-
- if (width) {
- ok &= gst_structure_get_int (structure, "width", width);
+ if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
+ blue_mask == 0xff000000) {
+ return GST_VIDEO_FORMAT_BGRx;
}
-
- if (height) {
- ok &= gst_structure_get_int (structure, "height", height);
+ if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
+ blue_mask == 0x000000ff) {
+ return GST_VIDEO_FORMAT_xRGB;
+ }
+ if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
+ blue_mask == 0x00ff0000) {
+ return GST_VIDEO_FORMAT_xBGR;
}
- return ok;
+ return GST_VIDEO_FORMAT_UNKNOWN;
}
-
-/**
- * gst_video_parse_caps_framerate:
- * @caps: pointer to a #GstCaps instance
- * @fps_n: pointer to integer to hold numerator of frame rate (output)
- * @fps_d: pointer to integer to hold denominator of frame rate (output)
- *
- * Extracts the frame rate from @caps and places the values in the locations
- * pointed to by @fps_n and @fps_d. Returns TRUE if the values could be
- * parsed correctly, FALSE if not.
- *
- * This function can be used with #GstCaps that have any media type; it
- * is not limited to formats handled by #GstVideoFormat.
- *
- * Since: 0.10.16
- *
- * Returns: TRUE if @caps was parsed correctly.
- */
-gboolean
-gst_video_parse_caps_framerate (GstCaps * caps, int *fps_n, int *fps_d)
+static GstVideoFormat
+gst_video_format_from_rgba32_masks (int red_mask, int green_mask,
+ int blue_mask, int alpha_mask)
{
- GstStructure *structure;
-
- if (!gst_caps_is_fixed (caps))
- return FALSE;
-
- structure = gst_caps_get_structure (caps, 0);
-
- return gst_structure_get_fraction (structure, "framerate", fps_n, fps_d);
+ if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
+ blue_mask == 0x0000ff00 && alpha_mask == 0x000000ff) {
+ return GST_VIDEO_FORMAT_RGBA;
+ }
+ if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
+ blue_mask == 0xff000000 && alpha_mask == 0x000000ff) {
+ return GST_VIDEO_FORMAT_BGRA;
+ }
+ if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
+ blue_mask == 0x000000ff && alpha_mask == 0xff000000) {
+ return GST_VIDEO_FORMAT_ARGB;
+ }
+ if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
+ blue_mask == 0x00ff0000 && alpha_mask == 0xff000000) {
+ return GST_VIDEO_FORMAT_ABGR;
+ }
+ return GST_VIDEO_FORMAT_UNKNOWN;
}
-/**
- * gst_video_parse_caps_pixel_aspect_ratio:
- * @caps: pointer to a #GstCaps instance
- * @par_n: pointer to numerator of pixel aspect ratio (output)
- * @par_d: pointer to denominator of pixel aspect ratio (output)
- *
- * Extracts the pixel aspect ratio from @caps and places the values in
- * the locations pointed to by @par_n and @par_d. Returns TRUE if the
- * values could be parsed correctly, FALSE if not.
- *
- * This function can be used with #GstCaps that have any media type; it
- * is not limited to formats handled by #GstVideoFormat.
- *
- * Since: 0.10.16
- *
- * Returns: TRUE if @caps was parsed correctly.
- */
-gboolean
-gst_video_parse_caps_pixel_aspect_ratio (GstCaps * caps, int *par_n, int *par_d)
+static GstVideoFormat
+gst_video_format_from_rgb24_masks (int red_mask, int green_mask, int blue_mask)
{
- GstStructure *structure;
-
- if (!gst_caps_is_fixed (caps))
- return FALSE;
-
- structure = gst_caps_get_structure (caps, 0);
-
- if (!gst_structure_get_fraction (structure, "pixel-aspect-ratio",
- par_n, par_d)) {
- *par_n = 1;
- *par_d = 1;
+ if (red_mask == 0xff0000 && green_mask == 0x00ff00 && blue_mask == 0x0000ff) {
+ return GST_VIDEO_FORMAT_RGB;
}
- return TRUE;
+ if (red_mask == 0x0000ff && green_mask == 0x00ff00 && blue_mask == 0xff0000) {
+ return GST_VIDEO_FORMAT_BGR;
+ }
+
+ return GST_VIDEO_FORMAT_UNKNOWN;
}
-/**
- * gst_video_format_new_caps_interlaced:
- * @format: the #GstVideoFormat describing the raw video format
- * @width: width of video
- * @height: height of video
- * @framerate_n: numerator of frame rate
- * @framerate_d: denominator of frame rate
- * @par_n: numerator of pixel aspect ratio
- * @par_d: denominator of pixel aspect ratio
- * @interlaced: #TRUE if the format is interlaced
- *
- * Creates a new #GstCaps object based on the parameters provided.
- *
- * Since: 0.10.23
- *
- * Returns: a new #GstCaps object, or NULL if there was an error
- */
-GstCaps *
-gst_video_format_new_caps_interlaced (GstVideoFormat format, int width,
- int height, int framerate_n, int framerate_d, int par_n, int par_d,
- gboolean interlaced)
-{
- GstCaps *res;
+#define GST_VIDEO_COMP1_MASK_16_INT 0xf800
+#define GST_VIDEO_COMP2_MASK_16_INT 0x07e0
+#define GST_VIDEO_COMP3_MASK_16_INT 0x001f
- res =
- gst_video_format_new_caps (format, width, height, framerate_n,
- framerate_d, par_n, par_d);
- if (interlaced && (res != NULL))
- gst_caps_set_simple (res, "interlaced", G_TYPE_BOOLEAN, TRUE, NULL);
+#define GST_VIDEO_COMP1_MASK_15_INT 0x7c00
+#define GST_VIDEO_COMP2_MASK_15_INT 0x03e0
+#define GST_VIDEO_COMP3_MASK_15_INT 0x001f
- return res;
+static GstVideoFormat
+gst_video_format_from_rgb16_masks (int red_mask, int green_mask, int blue_mask)
+{
+ if (red_mask == GST_VIDEO_COMP1_MASK_16_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_16_INT
+ && blue_mask == GST_VIDEO_COMP3_MASK_16_INT) {
+ return GST_VIDEO_FORMAT_RGB16;
+ }
+ if (red_mask == GST_VIDEO_COMP3_MASK_16_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_16_INT
+ && blue_mask == GST_VIDEO_COMP1_MASK_16_INT) {
+ return GST_VIDEO_FORMAT_BGR16;
+ }
+ if (red_mask == GST_VIDEO_COMP1_MASK_15_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_15_INT
+ && blue_mask == GST_VIDEO_COMP3_MASK_15_INT) {
+ return GST_VIDEO_FORMAT_RGB15;
+ }
+ if (red_mask == GST_VIDEO_COMP3_MASK_15_INT
+ && green_mask == GST_VIDEO_COMP2_MASK_15_INT
+ && blue_mask == GST_VIDEO_COMP1_MASK_15_INT) {
+ return GST_VIDEO_FORMAT_BGR15;
+ }
+ return GST_VIDEO_FORMAT_UNKNOWN;
}
-/**
- * gst_video_format_new_caps:
- * @format: the #GstVideoFormat describing the raw video format
- * @width: width of video
- * @height: height of video
- * @framerate_n: numerator of frame rate
- * @framerate_d: denominator of frame rate
- * @par_n: numerator of pixel aspect ratio
- * @par_d: denominator of pixel aspect ratio
- *
- * Creates a new #GstCaps object based on the parameters provided.
- *
- * Since: 0.10.16
- *
- * Returns: a new #GstCaps object, or NULL if there was an error
- */
-GstCaps *
-gst_video_format_new_caps (GstVideoFormat format, int width, int height,
- int framerate_n, int framerate_d, int par_n, int par_d)
+GstVideoFormat
+gst_video_format_from_masks (gint depth, gint bpp, gint endianness,
+ gint red_mask, gint green_mask, gint blue_mask, gint alpha_mask)
{
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
- g_return_val_if_fail (width > 0 && height > 0, NULL);
-
- if (gst_video_format_is_yuv (format)) {
- return gst_caps_new_simple ("video/x-raw-yuv",
- "format", GST_TYPE_FOURCC, gst_video_format_to_fourcc (format),
- "width", G_TYPE_INT, width,
- "height", G_TYPE_INT, height,
- "framerate", GST_TYPE_FRACTION, framerate_n, framerate_d,
- "pixel-aspect-ratio", GST_TYPE_FRACTION, par_n, par_d, NULL);
- }
- if (gst_video_format_is_rgb (format)) {
- GstCaps *caps;
- int red_mask;
- int blue_mask;
- int green_mask;
- int alpha_mask;
- int depth;
- int bpp;
- gboolean have_alpha;
- unsigned int mask;
-
- switch (format) {
- case GST_VIDEO_FORMAT_RGBx:
- case GST_VIDEO_FORMAT_BGRx:
- case GST_VIDEO_FORMAT_xRGB:
- case GST_VIDEO_FORMAT_xBGR:
- bpp = 32;
- depth = 24;
- have_alpha = FALSE;
- break;
- case GST_VIDEO_FORMAT_RGBA:
- case GST_VIDEO_FORMAT_BGRA:
- case GST_VIDEO_FORMAT_ARGB:
- case GST_VIDEO_FORMAT_ABGR:
- bpp = 32;
- depth = 32;
- have_alpha = TRUE;
- break;
- case GST_VIDEO_FORMAT_RGB:
- case GST_VIDEO_FORMAT_BGR:
- bpp = 24;
- depth = 24;
- have_alpha = FALSE;
- break;
- default:
- return NULL;
+ GstVideoFormat format;
+
+ /* our caps system handles 24/32bpp RGB as big-endian. */
+ if ((bpp == 24 || bpp == 32) && endianness == G_LITTLE_ENDIAN) {
+ red_mask = GUINT32_TO_BE (red_mask);
+ green_mask = GUINT32_TO_BE (green_mask);
+ blue_mask = GUINT32_TO_BE (blue_mask);
+ endianness = G_BIG_ENDIAN;
+ if (bpp == 24) {
+ red_mask >>= 8;
+ green_mask >>= 8;
+ blue_mask >>= 8;
}
- if (bpp == 32) {
- mask = 0xff000000;
+ }
+
+ if (depth == 30 && bpp == 32) {
+ format = GST_VIDEO_FORMAT_r210;
+ } else if (depth == 24 && bpp == 32) {
+ format = gst_video_format_from_rgb32_masks (red_mask, green_mask,
+ blue_mask);
+ } else if (depth == 32 && bpp == 32 && alpha_mask) {
+ format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
+ blue_mask, alpha_mask);
+ } else if (depth == 24 && bpp == 24) {
+ format = gst_video_format_from_rgb24_masks (red_mask, green_mask,
+ blue_mask);
+ } else if ((depth == 15 || depth == 16) && bpp == 16 &&
+ endianness == G_BYTE_ORDER) {
+ format = gst_video_format_from_rgb16_masks (red_mask, green_mask,
+ blue_mask);
+ } else if (depth == 8 && bpp == 8) {
+ format = GST_VIDEO_FORMAT_RGB8_PALETTED;
+ } else if (depth == 64 && bpp == 64) {
+ format = gst_video_format_from_rgba32_masks (red_mask, green_mask,
+ blue_mask, alpha_mask);
+ if (format == GST_VIDEO_FORMAT_ARGB) {
+ format = GST_VIDEO_FORMAT_ARGB64;
} else {
- mask = 0xff0000;
+ format = GST_VIDEO_FORMAT_UNKNOWN;
}
- red_mask =
- mask >> (8 * gst_video_format_get_component_offset (format, 0, width,
- height));
- green_mask =
- mask >> (8 * gst_video_format_get_component_offset (format, 1, width,
- height));
- blue_mask =
- mask >> (8 * gst_video_format_get_component_offset (format, 2, width,
- height));
-
- caps = gst_caps_new_simple ("video/x-raw-rgb",
- "bpp", G_TYPE_INT, bpp,
- "depth", G_TYPE_INT, depth,
- "endianness", G_TYPE_INT, G_BIG_ENDIAN,
- "red_mask", G_TYPE_INT, red_mask,
- "green_mask", G_TYPE_INT, green_mask,
- "blue_mask", G_TYPE_INT, blue_mask,
- "width", G_TYPE_INT, width,
- "height", G_TYPE_INT, height,
- "framerate", GST_TYPE_FRACTION, framerate_n, framerate_d,
- "pixel-aspect-ratio", GST_TYPE_FRACTION, par_n, par_d, NULL);
- if (have_alpha) {
- alpha_mask =
- mask >> (8 * gst_video_format_get_component_offset (format, 3, width,
- height));
- gst_caps_set_simple (caps, "alpha_mask", G_TYPE_INT, alpha_mask, NULL);
- }
- return caps;
+ } else {
+ format = GST_VIDEO_FORMAT_UNKNOWN;
}
- return NULL;
+ return format;
}
/**
return GST_VIDEO_FORMAT_YV12;
case GST_MAKE_FOURCC ('Y', 'U', 'Y', '2'):
return GST_VIDEO_FORMAT_YUY2;
+ case GST_MAKE_FOURCC ('Y', 'V', 'Y', 'U'):
+ return GST_VIDEO_FORMAT_YVYU;
case GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y'):
return GST_VIDEO_FORMAT_UYVY;
case GST_MAKE_FOURCC ('A', 'Y', 'U', 'V'):
return GST_VIDEO_FORMAT_Y41B;
case GST_MAKE_FOURCC ('Y', '4', '2', 'B'):
return GST_VIDEO_FORMAT_Y42B;
+ case GST_MAKE_FOURCC ('Y', '4', '4', '4'):
+ return GST_VIDEO_FORMAT_Y444;
+ case GST_MAKE_FOURCC ('v', '2', '1', '0'):
+ return GST_VIDEO_FORMAT_v210;
+ case GST_MAKE_FOURCC ('v', '2', '1', '6'):
+ return GST_VIDEO_FORMAT_v216;
+ case GST_MAKE_FOURCC ('N', 'V', '1', '2'):
+ return GST_VIDEO_FORMAT_NV12;
+ case GST_MAKE_FOURCC ('N', 'V', '2', '1'):
+ return GST_VIDEO_FORMAT_NV21;
+ case GST_MAKE_FOURCC ('v', '3', '0', '8'):
+ return GST_VIDEO_FORMAT_v308;
+ case GST_MAKE_FOURCC ('Y', '8', '0', '0'):
+ case GST_MAKE_FOURCC ('Y', '8', ' ', ' '):
+ case GST_MAKE_FOURCC ('G', 'R', 'E', 'Y'):
+ return GST_VIDEO_FORMAT_Y800;
+ case GST_MAKE_FOURCC ('Y', '1', '6', ' '):
+ return GST_VIDEO_FORMAT_Y16;
+ case GST_MAKE_FOURCC ('U', 'Y', 'V', 'P'):
+ return GST_VIDEO_FORMAT_UYVP;
+ case GST_MAKE_FOURCC ('A', '4', '2', '0'):
+ return GST_VIDEO_FORMAT_A420;
+ case GST_MAKE_FOURCC ('Y', 'U', 'V', '9'):
+ return GST_VIDEO_FORMAT_YUV9;
+ case GST_MAKE_FOURCC ('Y', 'V', 'U', '9'):
+ return GST_VIDEO_FORMAT_YVU9;
+ case GST_MAKE_FOURCC ('I', 'Y', 'U', '1'):
+ return GST_VIDEO_FORMAT_IYU1;
+ case GST_MAKE_FOURCC ('A', 'Y', '6', '4'):
+ return GST_VIDEO_FORMAT_AYUV64;
default:
return GST_VIDEO_FORMAT_UNKNOWN;
}
}
+GstVideoFormat
+gst_video_format_from_string (const gchar * format)
+{
+ guint i;
+
+ for (i = 0; i < G_N_ELEMENTS (formats); i++) {
+ if (strcmp (formats[i].fmt, format) == 0)
+ return formats[i].format;
+ }
+ return GST_VIDEO_FORMAT_UNKNOWN;
+}
+
+
/**
* gst_video_format_to_fourcc:
* @format: a #GstVideoFormat video format
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
- switch (format) {
- case GST_VIDEO_FORMAT_I420:
- return GST_MAKE_FOURCC ('I', '4', '2', '0');
- case GST_VIDEO_FORMAT_YV12:
- return GST_MAKE_FOURCC ('Y', 'V', '1', '2');
- case GST_VIDEO_FORMAT_YUY2:
- return GST_MAKE_FOURCC ('Y', 'U', 'Y', '2');
- case GST_VIDEO_FORMAT_UYVY:
- return GST_MAKE_FOURCC ('U', 'Y', 'V', 'Y');
- case GST_VIDEO_FORMAT_AYUV:
- return GST_MAKE_FOURCC ('A', 'Y', 'U', 'V');
- case GST_VIDEO_FORMAT_Y41B:
- return GST_MAKE_FOURCC ('Y', '4', '1', 'B');
- case GST_VIDEO_FORMAT_Y42B:
- return GST_MAKE_FOURCC ('Y', '4', '2', 'B');
- default:
- return 0;
- }
-}
-
-/*
- * gst_video_format_from_rgb32_masks:
- * @red_mask: red bit mask
- * @green_mask: green bit mask
- * @blue_mask: blue bit mask
- *
- * Converts red, green, blue bit masks into the corresponding
- * #GstVideoFormat.
- *
- * Since: 0.10.16
- *
- * Returns: the #GstVideoFormat corresponding to the bit masks
- */
-static GstVideoFormat
-gst_video_format_from_rgb32_masks (int red_mask, int green_mask, int blue_mask)
-{
- if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
- blue_mask == 0x0000ff00) {
- return GST_VIDEO_FORMAT_RGBx;
- }
- if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
- blue_mask == 0xff000000) {
- return GST_VIDEO_FORMAT_BGRx;
- }
- if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
- blue_mask == 0x000000ff) {
- return GST_VIDEO_FORMAT_xRGB;
- }
- if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
- blue_mask == 0x00ff0000) {
- return GST_VIDEO_FORMAT_xBGR;
- }
+ if (format >= G_N_ELEMENTS (formats))
+ return 0;
- return GST_VIDEO_FORMAT_UNKNOWN;
+ return formats[format].fourcc;
}
-static GstVideoFormat
-gst_video_format_from_rgba32_masks (int red_mask, int green_mask, int blue_mask,
- int alpha_mask)
+const gchar *
+gst_video_format_to_string (GstVideoFormat format)
{
- if (red_mask == 0xff000000 && green_mask == 0x00ff0000 &&
- blue_mask == 0x0000ff00 && alpha_mask == 0x000000ff) {
- return GST_VIDEO_FORMAT_RGBA;
- }
- if (red_mask == 0x0000ff00 && green_mask == 0x00ff0000 &&
- blue_mask == 0xff000000 && alpha_mask == 0x000000ff) {
- return GST_VIDEO_FORMAT_BGRA;
- }
- if (red_mask == 0x00ff0000 && green_mask == 0x0000ff00 &&
- blue_mask == 0x000000ff && alpha_mask == 0xff000000) {
- return GST_VIDEO_FORMAT_ARGB;
- }
- if (red_mask == 0x000000ff && green_mask == 0x0000ff00 &&
- blue_mask == 0x00ff0000 && alpha_mask == 0xff000000) {
- return GST_VIDEO_FORMAT_ABGR;
- }
-
- return GST_VIDEO_FORMAT_UNKNOWN;
-}
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
-static GstVideoFormat
-gst_video_format_from_rgb24_masks (int red_mask, int green_mask, int blue_mask)
-{
- if (red_mask == 0xff0000 && green_mask == 0x00ff00 && blue_mask == 0x0000ff) {
- return GST_VIDEO_FORMAT_RGB;
- }
- if (red_mask == 0x0000ff && green_mask == 0x00ff00 && blue_mask == 0xff0000) {
- return GST_VIDEO_FORMAT_BGR;
- }
+ if (format >= G_N_ELEMENTS (formats))
+ return NULL;
- return GST_VIDEO_FORMAT_UNKNOWN;
+ return formats[format].fmt;
}
/**
gboolean
gst_video_format_is_rgb (GstVideoFormat format)
{
- switch (format) {
- case GST_VIDEO_FORMAT_I420:
- case GST_VIDEO_FORMAT_YV12:
- case GST_VIDEO_FORMAT_YUY2:
- case GST_VIDEO_FORMAT_UYVY:
- case GST_VIDEO_FORMAT_AYUV:
- case GST_VIDEO_FORMAT_Y41B:
- case GST_VIDEO_FORMAT_Y42B:
- return FALSE;
- case GST_VIDEO_FORMAT_RGBx:
- case GST_VIDEO_FORMAT_BGRx:
- case GST_VIDEO_FORMAT_xRGB:
- case GST_VIDEO_FORMAT_xBGR:
- case GST_VIDEO_FORMAT_RGBA:
- case GST_VIDEO_FORMAT_BGRA:
- case GST_VIDEO_FORMAT_ARGB:
- case GST_VIDEO_FORMAT_ABGR:
- case GST_VIDEO_FORMAT_RGB:
- case GST_VIDEO_FORMAT_BGR:
- return TRUE;
- default:
- return FALSE;
- }
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return FALSE;
+
+ return (formats[format].flags & VIDEO_FLAG_RGB) != 0;
}
/**
gboolean
gst_video_format_is_yuv (GstVideoFormat format)
{
- switch (format) {
- case GST_VIDEO_FORMAT_I420:
- case GST_VIDEO_FORMAT_YV12:
- case GST_VIDEO_FORMAT_YUY2:
- case GST_VIDEO_FORMAT_UYVY:
- case GST_VIDEO_FORMAT_AYUV:
- case GST_VIDEO_FORMAT_Y41B:
- case GST_VIDEO_FORMAT_Y42B:
- return TRUE;
- case GST_VIDEO_FORMAT_RGBx:
- case GST_VIDEO_FORMAT_BGRx:
- case GST_VIDEO_FORMAT_xRGB:
- case GST_VIDEO_FORMAT_xBGR:
- case GST_VIDEO_FORMAT_RGBA:
- case GST_VIDEO_FORMAT_BGRA:
- case GST_VIDEO_FORMAT_ARGB:
- case GST_VIDEO_FORMAT_ABGR:
- case GST_VIDEO_FORMAT_RGB:
- case GST_VIDEO_FORMAT_BGR:
- return FALSE;
- default:
- return FALSE;
- }
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return FALSE;
+
+ return (formats[format].flags & VIDEO_FLAG_YUV) != 0;
+}
+
+/**
+ * gst_video_format_is_gray:
+ * @format: a #GstVideoFormat
+ *
+ * Determine whether the video format is a grayscale format.
+ *
+ * Since: 0.10.29
+ *
+ * Returns: TRUE if @format represents grayscale video
+ */
+gboolean
+gst_video_format_is_gray (GstVideoFormat format)
+{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return FALSE;
+
+ return (formats[format].flags & VIDEO_FLAG_GRAY) != 0;
}
/**
* gst_video_format_has_alpha:
* @format: a #GstVideoFormat
- *
+ *
* Returns TRUE or FALSE depending on if the video format provides an
* alpha channel.
*
gboolean
gst_video_format_has_alpha (GstVideoFormat format)
{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, FALSE);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return FALSE;
+
+ return (formats[format].flags & VIDEO_FLAG_ALPHA) != 0;
+}
+
+/**
+ * gst_video_format_get_n_components:
+ * @format: a #GstVideoFormat
+ *
+ * Get the number of components for @format.
+ *
+ * Returns: the number of components for @format.
+ */
+int
+gst_video_format_get_n_components (GstVideoFormat format)
+{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return 0;
+
+ return formats[format].n_comp;
+}
+
+/**
+ * gst_video_format_get_component_depth:
+ * @format: a #GstVideoFormat
+ * @component: the video component (e.g. 0 for 'R' in RGB)
+ *
+ * Returns the number of bits used to encode an individual pixel of
+ * a given @component. Typically this is 8, although higher and lower
+ * values are possible for some formats.
+ *
+ * Since: 0.10.33
+ *
+ * Returns: depth of component
+ */
+int
+gst_video_format_get_component_depth (GstVideoFormat format, int component)
+{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
+ g_return_val_if_fail (component < GST_VIDEO_MAX_PLANES, 0);
+
+ if (format >= G_N_ELEMENTS (formats))
+ return FALSE;
+
+ return formats[format].depth[component];
+}
+
+/**
+ * gst_video_format_get_pixel_stride:
+ * @format: a #GstVideoFormat
+ * @component: the component index
+ *
+ * Calculates the pixel stride (number of bytes from one pixel to the
+ * pixel to its immediate left) for the video component with an index
+ * of @component. See @gst_video_format_get_row_stride for a description
+ * of the component index.
+ *
+ * Since: 0.10.16
+ *
+ * Returns: pixel stride of component @component
+ */
+int
+gst_video_format_get_pixel_stride (GstVideoFormat format, int component)
+{
+ g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
+ g_return_val_if_fail (component >= 0 && component <= 3, 0);
+
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
- case GST_VIDEO_FORMAT_YUY2:
- case GST_VIDEO_FORMAT_UYVY:
case GST_VIDEO_FORMAT_Y41B:
case GST_VIDEO_FORMAT_Y42B:
- return FALSE;
+ case GST_VIDEO_FORMAT_Y444:
+ case GST_VIDEO_FORMAT_A420:
+ case GST_VIDEO_FORMAT_YUV9:
+ case GST_VIDEO_FORMAT_YVU9:
+ return 1;
+ case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
+ case GST_VIDEO_FORMAT_UYVY:
+ if (component == 0) {
+ return 2;
+ } else {
+ return 4;
+ }
+ case GST_VIDEO_FORMAT_IYU1:
+ /* doesn't make much sense for IYU1 because it's 1 or 3
+ * for luma depending on position */
+ return 0;
case GST_VIDEO_FORMAT_AYUV:
- case GST_VIDEO_FORMAT_RGBA:
- case GST_VIDEO_FORMAT_BGRA:
- case GST_VIDEO_FORMAT_ARGB:
- case GST_VIDEO_FORMAT_ABGR:
- return TRUE;
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_xBGR:
+ case GST_VIDEO_FORMAT_RGBA:
+ case GST_VIDEO_FORMAT_BGRA:
+ case GST_VIDEO_FORMAT_ARGB:
+ case GST_VIDEO_FORMAT_ABGR:
+ case GST_VIDEO_FORMAT_r210:
+ return 4;
+ case GST_VIDEO_FORMAT_RGB16:
+ case GST_VIDEO_FORMAT_BGR16:
+ case GST_VIDEO_FORMAT_RGB15:
+ case GST_VIDEO_FORMAT_BGR15:
+ return 2;
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
- return FALSE;
+ case GST_VIDEO_FORMAT_v308:
+ return 3;
+ case GST_VIDEO_FORMAT_v210:
+ /* v210 is packed at the bit level, so pixel stride doesn't make sense */
+ return 0;
+ case GST_VIDEO_FORMAT_v216:
+ if (component == 0) {
+ return 4;
+ } else {
+ return 8;
+ }
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
+ if (component == 0) {
+ return 1;
+ } else {
+ return 2;
+ }
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_Y800:
+ return 1;
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y16:
+ return 2;
+ case GST_VIDEO_FORMAT_UYVP:
+ /* UYVP is packed at the bit level, so pixel stride doesn't make sense */
+ return 0;
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
+ return 1;
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ return 8;
default:
- return FALSE;
+ return 0;
+ }
+}
+
+/**
+ * gst_video_info_init:
+ * @info: a #GstVideoInfo
+ *
+ * Initialize @info with default values.
+ */
+void
+gst_video_info_init (GstVideoInfo * info)
+{
+ g_return_if_fail (info != NULL);
+
+ memset (info, 0, sizeof (GstVideoInfo));
+}
+
+/**
+ * gst_video_info_set_format:
+ * @info: a #GstVideoInfo
+ * @format: the format
+ * @width: a width
+ * @height: a height
+ *
+ * Set the default info for a video frame of @format and @width and @height.
+ */
+void
+gst_video_info_set_format (GstVideoInfo * info, GstVideoFormat format,
+ guint width, guint height)
+{
+ gint i;
+
+ g_return_if_fail (info != NULL);
+ g_return_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN);
+
+ info->format = format;
+ info->width = width;
+ info->height = height;
+
+ info->n_planes = formats[format].n_planes;
+ info->size = get_size (format, info->width, info->height);
+ fill_planes (info);
+
+ for (i = 0; i < info->n_planes; i++) {
+ info->plane[i].stride = get_stride (format, i, info->width);
+ }
+}
+
+/**
+ * gst_video_info_from_caps:
+ * @info: a #GstVideoInfo
+ * @caps: a #GstCaps
+ *
+ * Parse @caps and update @info.
+ *
+ * Returns: TRUE if @caps could be parsed
+ */
+gboolean
+gst_video_info_from_caps (GstVideoInfo * info, const GstCaps * caps)
+{
+ GstStructure *structure;
+ const gchar *s;
+ GstVideoFormat format;
+ gint width, height;
+ gint fps_n, fps_d;
+ gboolean interlaced;
+ gint par_n, par_d;
+
+ g_return_val_if_fail (info != NULL, FALSE);
+ g_return_val_if_fail (caps != NULL, FALSE);
+ g_return_val_if_fail (gst_caps_is_fixed (caps), FALSE);
+
+ structure = gst_caps_get_structure (caps, 0);
+
+ if (!gst_structure_has_name (structure, "video/x-raw"))
+ goto wrong_name;
+
+ if (!(s = gst_structure_get_string (structure, "format")))
+ goto no_format;
+
+ format = gst_video_format_from_string (s);
+ if (format == GST_VIDEO_FORMAT_UNKNOWN)
+ goto unknown_format;
+
+ if (!gst_structure_get_int (structure, "width", &width))
+ goto no_width;
+ if (!gst_structure_get_int (structure, "height", &height))
+ goto no_height;
+
+ gst_video_info_set_format (info, format, width, height);
+
+ if (gst_structure_get_fraction (structure, "framerate", &fps_n, &fps_d)) {
+ info->fps_n = fps_n;
+ info->fps_d = fps_d;
+ } else {
+ info->fps_n = 0;
+ info->fps_d = 1;
+ }
+
+ if (!gst_structure_get_boolean (structure, "interlaced", &interlaced))
+ interlaced = FALSE;
+ if (interlaced)
+ info->flags |= GST_VIDEO_FLAG_INTERLACED;
+ else
+ info->flags &= ~GST_VIDEO_FLAG_INTERLACED;
+
+ s = gst_structure_get_string (structure, "color-matrix");
+ if (s)
+ info->color_matrix = s;
+ else
+ info->color_matrix = "sdtv";
+
+ s = gst_structure_get_string (structure, "chroma-site");
+ if (s)
+ info->chroma_site = s;
+ else
+ info->chroma_site = "mpeg2";
+
+ if (gst_structure_get_fraction (structure, "pixel-aspect-ratio",
+ &par_n, &par_d)) {
+ info->par_n = par_n;
+ info->par_d = par_d;
+ } else {
+ info->par_n = 1;
+ info->par_d = 1;
+ }
+ return TRUE;
+
+ /* ERROR */
+wrong_name:
+ {
+ return FALSE;
+ }
+no_format:
+ {
+ return FALSE;
+ }
+unknown_format:
+ {
+ return FALSE;
+ }
+no_width:
+ {
+ return FALSE;
+ }
+no_height:
+ {
+ return FALSE;
}
}
/**
- * gst_video_format_get_row_stride:
+ * gst_video_info_to_caps:
+ * @info: a #GstVideoInfo
+ *
+ * Convert the values of @info into a #GstCaps.
+ *
+ * Returns: a new #GstCaps containing the info of @info.
+ */
+GstCaps *
+gst_video_info_to_caps (GstVideoInfo * info)
+{
+ GstCaps *caps;
+ const gchar *format;
+
+ g_return_val_if_fail (info != NULL, NULL);
+ g_return_val_if_fail (info->format != GST_VIDEO_FORMAT_UNKNOWN, NULL);
+
+ format = gst_video_format_to_string (info->format);
+ g_return_val_if_fail (format != NULL, NULL);
+
+ caps = gst_caps_new_simple ("video/x-raw",
+ "format", G_TYPE_STRING, format,
+ "width", G_TYPE_INT, info->width,
+ "height", G_TYPE_INT, info->height,
+ "framerate", GST_TYPE_FRACTION, info->fps_n, info->fps_d,
+ "pixel-aspect-ratio", GST_TYPE_FRACTION, info->par_n, info->par_d, NULL);
+ if (info->flags & GST_VIDEO_FLAG_INTERLACED)
+ gst_caps_set_simple (caps, "interlaced", G_TYPE_BOOLEAN, TRUE, NULL);
+ if (info->color_matrix)
+ gst_caps_set_simple (caps, "color-matrix", G_TYPE_STRING,
+ info->color_matrix, NULL);
+ if (info->chroma_site)
+ gst_caps_set_simple (caps, "chromar-site", G_TYPE_STRING, info->chroma_site,
+ NULL);
+
+ return caps;
+}
+
+/**
+ * get_stride:
* @format: a #GstVideoFormat
* @component: the component index
* @width: the width of video
*
* Returns: row stride of component @component
*/
-int
-gst_video_format_get_row_stride (GstVideoFormat format, int component,
- int width)
+static int
+get_stride (GstVideoFormat format, int plane, int width)
{
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
- g_return_val_if_fail (component >= 0 && component <= 3, 0);
- g_return_val_if_fail (width > 0, 0);
-
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
- if (component == 0) {
+ if (plane == 0) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
}
case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
return GST_ROUND_UP_4 (width * 2);
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
+ case GST_VIDEO_FORMAT_r210:
return width * 4;
+ case GST_VIDEO_FORMAT_RGB16:
+ case GST_VIDEO_FORMAT_BGR16:
+ case GST_VIDEO_FORMAT_RGB15:
+ case GST_VIDEO_FORMAT_BGR15:
+ return GST_ROUND_UP_4 (width * 2);
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
+ case GST_VIDEO_FORMAT_v308:
return GST_ROUND_UP_4 (width * 3);
case GST_VIDEO_FORMAT_Y41B:
- if (component == 0) {
+ if (plane == 0) {
return GST_ROUND_UP_4 (width);
} else {
- return GST_ROUND_UP_8 (width) / 4;
+ return GST_ROUND_UP_16 (width) / 4;
}
case GST_VIDEO_FORMAT_Y42B:
- if (component == 0) {
+ if (plane == 0) {
return GST_ROUND_UP_4 (width);
} else {
return GST_ROUND_UP_8 (width) / 2;
}
- default:
- return 0;
- }
-}
-
-/**
- * gst_video_format_get_pixel_stride:
- * @format: a #GstVideoFormat
- * @component: the component index
- *
- * Calculates the pixel stride (number of bytes from one pixel to the
- * pixel to its immediate left) for the video component with an index
- * of @component. See @gst_video_format_get_row_stride for a description
- * of the component index.
- *
- * Since: 0.10.16
- *
- * Returns: pixel stride of component @component
- */
-int
-gst_video_format_get_pixel_stride (GstVideoFormat format, int component)
-{
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
- g_return_val_if_fail (component >= 0 && component <= 3, 0);
-
- switch (format) {
- case GST_VIDEO_FORMAT_I420:
- case GST_VIDEO_FORMAT_YV12:
- case GST_VIDEO_FORMAT_Y41B:
- case GST_VIDEO_FORMAT_Y42B:
- return 1;
- case GST_VIDEO_FORMAT_YUY2:
- case GST_VIDEO_FORMAT_UYVY:
- if (component == 0) {
- return 2;
+ case GST_VIDEO_FORMAT_Y444:
+ return GST_ROUND_UP_4 (width);
+ case GST_VIDEO_FORMAT_v210:
+ return ((width + 47) / 48) * 128;
+ case GST_VIDEO_FORMAT_v216:
+ return GST_ROUND_UP_8 (width * 4);
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
+ return GST_ROUND_UP_4 (width);
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_Y800:
+ return GST_ROUND_UP_4 (width);
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y16:
+ return GST_ROUND_UP_4 (width * 2);
+ case GST_VIDEO_FORMAT_UYVP:
+ return GST_ROUND_UP_4 ((width * 2 * 5 + 3) / 4);
+ case GST_VIDEO_FORMAT_A420:
+ if (plane == 0 || plane == 3) {
+ return GST_ROUND_UP_4 (width);
} else {
- return 4;
+ return GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2);
}
- case GST_VIDEO_FORMAT_AYUV:
- case GST_VIDEO_FORMAT_RGBx:
- case GST_VIDEO_FORMAT_BGRx:
- case GST_VIDEO_FORMAT_xRGB:
- case GST_VIDEO_FORMAT_xBGR:
- case GST_VIDEO_FORMAT_RGBA:
- case GST_VIDEO_FORMAT_BGRA:
- case GST_VIDEO_FORMAT_ARGB:
- case GST_VIDEO_FORMAT_ABGR:
- return 4;
- case GST_VIDEO_FORMAT_RGB:
- case GST_VIDEO_FORMAT_BGR:
- return 3;
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
+ return GST_ROUND_UP_4 (width);
+ case GST_VIDEO_FORMAT_YUV9:
+ case GST_VIDEO_FORMAT_YVU9:
+ if (plane == 0) {
+ return GST_ROUND_UP_4 (width);
+ } else {
+ return GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4);
+ }
+ case GST_VIDEO_FORMAT_IYU1:
+ return GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) +
+ GST_ROUND_UP_4 (width) / 2);
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ return width * 8;
default:
return 0;
}
}
+#if 0
/**
* gst_video_format_get_component_width:
* @format: a #GstVideoFormat
*
* Returns: width of component @component
*/
-int
-gst_video_format_get_component_width (GstVideoFormat format, int component,
- int width)
+static int
+gst_video_format_get_component_width (GstVideoFormat format,
+ int component, int width)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
+ case GST_VIDEO_FORMAT_Y42B:
+ case GST_VIDEO_FORMAT_v210:
+ case GST_VIDEO_FORMAT_v216:
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
+ case GST_VIDEO_FORMAT_UYVP:
if (component == 0) {
return width;
} else {
return GST_ROUND_UP_2 (width) / 2;
}
- case GST_VIDEO_FORMAT_Y41B: /* CHECKME: component_width for Y41B */
- if (component == 0) {
- return width;
- } else {
- return GST_ROUND_UP_8 (width) / 4;
- }
- case GST_VIDEO_FORMAT_Y42B: /* CHECKME: component_width for Y42B */
+ case GST_VIDEO_FORMAT_Y41B:
+ case GST_VIDEO_FORMAT_YUV9:
+ case GST_VIDEO_FORMAT_YVU9:
+ case GST_VIDEO_FORMAT_IYU1:
if (component == 0) {
return width;
} else {
- return GST_ROUND_UP_8 (width) / 2;
+ return GST_ROUND_UP_4 (width) / 4;
}
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_ABGR:
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
+ case GST_VIDEO_FORMAT_RGB16:
+ case GST_VIDEO_FORMAT_BGR16:
+ case GST_VIDEO_FORMAT_RGB15:
+ case GST_VIDEO_FORMAT_BGR15:
+ case GST_VIDEO_FORMAT_Y444:
+ case GST_VIDEO_FORMAT_v308:
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y800:
+ case GST_VIDEO_FORMAT_Y16:
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ case GST_VIDEO_FORMAT_r210:
return width;
+ case GST_VIDEO_FORMAT_A420:
+ if (component == 0 || component == 3) {
+ return width;
+ } else {
+ return GST_ROUND_UP_2 (width) / 2;
+ }
default:
return 0;
}
}
+#endif
+#if 0
/**
* gst_video_format_get_component_height:
* @format: a #GstVideoFormat
* Returns: height of component @component
*/
int
-gst_video_format_get_component_height (GstVideoFormat format, int component,
- int height)
+gst_video_format_get_component_height (GstVideoFormat format,
+ int component, int height)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
case GST_VIDEO_FORMAT_YV12:
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
if (component == 0) {
return height;
} else {
case GST_VIDEO_FORMAT_Y41B:
case GST_VIDEO_FORMAT_Y42B:
case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_ABGR:
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
+ case GST_VIDEO_FORMAT_RGB16:
+ case GST_VIDEO_FORMAT_BGR16:
+ case GST_VIDEO_FORMAT_RGB15:
+ case GST_VIDEO_FORMAT_BGR15:
+ case GST_VIDEO_FORMAT_Y444:
+ case GST_VIDEO_FORMAT_v210:
+ case GST_VIDEO_FORMAT_v216:
+ case GST_VIDEO_FORMAT_v308:
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y800:
+ case GST_VIDEO_FORMAT_Y16:
+ case GST_VIDEO_FORMAT_UYVP:
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
+ case GST_VIDEO_FORMAT_IYU1:
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ case GST_VIDEO_FORMAT_r210:
return height;
+ case GST_VIDEO_FORMAT_A420:
+ if (component == 0 || component == 3) {
+ return height;
+ } else {
+ return GST_ROUND_UP_2 (height) / 2;
+ }
+ case GST_VIDEO_FORMAT_YUV9:
+ case GST_VIDEO_FORMAT_YVU9:
+ if (component == 0) {
+ return height;
+ } else {
+ return GST_ROUND_UP_4 (height) / 4;
+ }
default:
return 0;
}
}
+#endif
+#if 0
/**
- * gst_video_format_get_component_offset:
+ * get_component_offset:
* @format: a #GstVideoFormat
* @component: the component index
* @width: the width of video
*
* Returns: offset of component @component
*/
-int
-gst_video_format_get_component_offset (GstVideoFormat format, int component,
- int width, int height)
+static int
+get_component_offset (GstVideoFormat format,
+ int component, int width, int height)
{
g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
g_return_val_if_fail (component >= 0 && component <= 3, 0);
- g_return_val_if_fail (width > 0 && height > 0, 0);
+ g_return_val_if_fail ((!gst_video_format_is_yuv (format)) || (width > 0
+ && height > 0), 0);
switch (format) {
case GST_VIDEO_FORMAT_I420:
GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2);
}
- return 0;
+ break;
case GST_VIDEO_FORMAT_YV12: /* same as I420, but components 1+2 swapped */
if (component == 0)
return 0;
GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2);
}
- return 0;
+ break;
case GST_VIDEO_FORMAT_YUY2:
if (component == 0)
return 0;
return 1;
if (component == 2)
return 3;
- return 0;
+ break;
+ case GST_VIDEO_FORMAT_YVYU:
+ if (component == 0)
+ return 0;
+ if (component == 1)
+ return 3;
+ if (component == 2)
+ return 1;
+ break;
case GST_VIDEO_FORMAT_UYVY:
if (component == 0)
return 1;
return 0;
if (component == 2)
return 2;
- return 0;
+ break;
case GST_VIDEO_FORMAT_AYUV:
if (component == 0)
return 1;
return 3;
if (component == 3)
return 0;
- return 0;
+ break;
case GST_VIDEO_FORMAT_RGBx:
case GST_VIDEO_FORMAT_RGBA:
if (component == 0)
return 2;
if (component == 3)
return 3;
- return 0;
+ break;
case GST_VIDEO_FORMAT_BGRx:
case GST_VIDEO_FORMAT_BGRA:
if (component == 0)
return 0;
if (component == 3)
return 3;
- return 0;
+ break;
case GST_VIDEO_FORMAT_xRGB:
case GST_VIDEO_FORMAT_ARGB:
if (component == 0)
return 3;
if (component == 3)
return 0;
- return 0;
+ break;
case GST_VIDEO_FORMAT_xBGR:
case GST_VIDEO_FORMAT_ABGR:
if (component == 0)
return 1;
if (component == 3)
return 0;
- return 0;
+ break;
case GST_VIDEO_FORMAT_RGB:
+ case GST_VIDEO_FORMAT_v308:
if (component == 0)
return 0;
if (component == 1)
return 1;
if (component == 2)
return 2;
- return 0;
+ break;
case GST_VIDEO_FORMAT_BGR:
if (component == 0)
return 2;
return 1;
if (component == 2)
return 0;
- return 0;
+ break;
case GST_VIDEO_FORMAT_Y41B:
if (component == 0)
return 0;
if (component == 1)
return GST_ROUND_UP_4 (width) * height;
if (component == 2)
- return (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_8 (width) / 4)) * height;
- return 0;
+ return (GST_ROUND_UP_4 (width) +
+ (GST_ROUND_UP_16 (width) / 4)) * height;
+ break;
case GST_VIDEO_FORMAT_Y42B:
if (component == 0)
return 0;
return GST_ROUND_UP_4 (width) * height;
if (component == 2)
return (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_8 (width) / 2)) * height;
+ break;
+ case GST_VIDEO_FORMAT_Y444:
+ return GST_ROUND_UP_4 (width) * height * component;
+ case GST_VIDEO_FORMAT_v210:
+ case GST_VIDEO_FORMAT_r210:
+ /* v210 is bit-packed, so this doesn't make sense */
return 0;
- default:
+ case GST_VIDEO_FORMAT_v216:
+ if (component == 0)
+ return 0;
+ if (component == 1)
+ return 2;
+ if (component == 2)
+ return 6;
+ break;
+ case GST_VIDEO_FORMAT_NV12:
+ if (component == 0)
+ return 0;
+ if (component == 1)
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ if (component == 2)
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) + 1;
+ break;
+ case GST_VIDEO_FORMAT_NV21:
+ if (component == 0)
+ return 0;
+ if (component == 1)
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) + 1;
+ if (component == 2)
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ break;
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y800:
+ case GST_VIDEO_FORMAT_Y16:
+ return 0;
+ case GST_VIDEO_FORMAT_UYVP:
+ /* UYVP is bit-packed, so this doesn't make sense */
+ return 0;
+ case GST_VIDEO_FORMAT_A420:
+ if (component == 0)
+ return 0;
+ if (component == 1)
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ if (component == 2) {
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) +
+ GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2);
+ }
+ if (component == 3) {
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) +
+ 2 * GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2);
+ }
+ break;
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
return 0;
+ case GST_VIDEO_FORMAT_YUV9:
+ if (component == 0)
+ return 0;
+ if (component == 1)
+ return GST_ROUND_UP_4 (width) * height;
+ if (component == 2) {
+ return GST_ROUND_UP_4 (width) * height +
+ GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
+ (GST_ROUND_UP_4 (height) / 4);
+ }
+ break;
+ case GST_VIDEO_FORMAT_YVU9:
+ if (component == 0)
+ return 0;
+ if (component == 1) {
+ return GST_ROUND_UP_4 (width) * height +
+ GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
+ (GST_ROUND_UP_4 (height) / 4);
+ }
+ if (component == 2)
+ return GST_ROUND_UP_4 (width) * height;
+ break;
+ case GST_VIDEO_FORMAT_IYU1:
+ if (component == 0)
+ return 1;
+ if (component == 1)
+ return 0;
+ if (component == 2)
+ return 4;
+ break;
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ if (component == 0)
+ return 2;
+ if (component == 1)
+ return 4;
+ if (component == 2)
+ return 6;
+ if (component == 3)
+ return 0;
+ break;
+ default:
+ break;
}
+ GST_WARNING ("unhandled format %d or component %d", format, component);
+ return 0;
}
+#endif
/**
- * gst_video_format_get_size:
+ * get_plane_offset:
* @format: a #GstVideoFormat
+ * @plane: the plane index
* @width: the width of video
* @height: the height of video
*
- * Calculates the total number of bytes in the raw video format. This
- * number should be used when allocating a buffer for raw video.
+ * Calculates the offset (in bytes) of the first pixel of the plane
+ * with index @plane. For packed formats, this will typically be 0.
+ * For planar formats, this will be a (relatively) large offset to the
+ * beginning of the second or third plane planes.
+ * See @gst_video_format_get_row_stride for a description
+ * of the plane index.
*
* Since: 0.10.16
*
+ * Returns: offset of plane @plane
+ */
+static int
+fill_planes (GstVideoInfo * info)
+{
+ int width;
+ int height;
+
+ width = info->width;
+ height = info->height;
+
+ switch (info->format) {
+ case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
+ case GST_VIDEO_FORMAT_UYVY:
+ case GST_VIDEO_FORMAT_AYUV:
+ case GST_VIDEO_FORMAT_RGBx:
+ case GST_VIDEO_FORMAT_RGBA:
+ case GST_VIDEO_FORMAT_BGRx:
+ case GST_VIDEO_FORMAT_BGRA:
+ case GST_VIDEO_FORMAT_xRGB:
+ case GST_VIDEO_FORMAT_ARGB:
+ case GST_VIDEO_FORMAT_xBGR:
+ case GST_VIDEO_FORMAT_ABGR:
+ case GST_VIDEO_FORMAT_RGB:
+ case GST_VIDEO_FORMAT_v308:
+ case GST_VIDEO_FORMAT_BGR:
+ case GST_VIDEO_FORMAT_v210:
+ case GST_VIDEO_FORMAT_r210:
+ case GST_VIDEO_FORMAT_v216:
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y800:
+ case GST_VIDEO_FORMAT_Y16:
+ case GST_VIDEO_FORMAT_UYVP:
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
+ case GST_VIDEO_FORMAT_IYU1:
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ info->n_planes = 1;
+ info->plane[0].offset = 0;
+ break;
+ case GST_VIDEO_FORMAT_I420:
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ info->plane[2].offset = info->plane[1].offset +
+ GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2);
+ break;
+ case GST_VIDEO_FORMAT_YV12: /* same as I420, but plane 1+2 swapped */
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[2].offset = GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ info->plane[1].offset = info->plane[2].offset +
+ GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2);
+ break;
+ case GST_VIDEO_FORMAT_Y41B:
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * height;
+ info->plane[2].offset = (GST_ROUND_UP_4 (width) +
+ (GST_ROUND_UP_16 (width) / 4)) * height;
+ break;
+ case GST_VIDEO_FORMAT_Y42B:
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * height;
+ info->plane[2].offset =
+ (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_8 (width) / 2)) * height;
+ break;
+ case GST_VIDEO_FORMAT_Y444:
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * height;
+ info->plane[2].offset = GST_ROUND_UP_4 (width) * height * 2;
+ break;
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
+ info->n_planes = 2;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ break;
+ case GST_VIDEO_FORMAT_A420:
+ info->n_planes = 4;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ info->plane[2].offset = info->plane[1].offset +
+ GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2);
+ info->plane[3].offset = info->plane[2].offset +
+ GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2);
+ break;
+ case GST_VIDEO_FORMAT_YUV9:
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[1].offset = GST_ROUND_UP_4 (width) * height;
+ info->plane[2].offset = info->plane[1].offset +
+ GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
+ (GST_ROUND_UP_4 (height) / 4);
+ break;
+ case GST_VIDEO_FORMAT_YVU9:
+ info->n_planes = 3;
+ info->plane[0].offset = 0;
+ info->plane[2].offset = GST_ROUND_UP_4 (width) * height;
+ info->plane[1].offset = info->plane[2].offset +
+ GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
+ (GST_ROUND_UP_4 (height) / 4);
+ break;
+ default:
+ break;
+ }
+ GST_WARNING ("unhandled format %d", info->format);
+ return 0;
+}
+
+/**
+ * get_size:
+ * @format: a #GstVideoFormat
+ * @width: the width of video
+ * @height: the height of video
+ *
+ * Calculates the total number of bytes in the raw video format. This
+ * number should be used when allocating a buffer for raw video.
+ *
* Returns: size (in bytes) of raw video format
*/
-int
-gst_video_format_get_size (GstVideoFormat format, int width, int height)
+static int
+get_size (GstVideoFormat format, int width, int height)
{
int size;
size += GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
(GST_ROUND_UP_2 (height) / 2) * 2;
return size;
+ case GST_VIDEO_FORMAT_IYU1:
+ return GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) +
+ GST_ROUND_UP_4 (width) / 2) * height;
case GST_VIDEO_FORMAT_YUY2:
+ case GST_VIDEO_FORMAT_YVYU:
case GST_VIDEO_FORMAT_UYVY:
return GST_ROUND_UP_4 (width * 2) * height;
case GST_VIDEO_FORMAT_AYUV:
case GST_VIDEO_FORMAT_BGRA:
case GST_VIDEO_FORMAT_ARGB:
case GST_VIDEO_FORMAT_ABGR:
+ case GST_VIDEO_FORMAT_r210:
return width * 4 * height;
+ case GST_VIDEO_FORMAT_RGB16:
+ case GST_VIDEO_FORMAT_BGR16:
+ case GST_VIDEO_FORMAT_RGB15:
+ case GST_VIDEO_FORMAT_BGR15:
+ return GST_ROUND_UP_4 (width * 2) * height;
case GST_VIDEO_FORMAT_RGB:
case GST_VIDEO_FORMAT_BGR:
+ case GST_VIDEO_FORMAT_v308:
return GST_ROUND_UP_4 (width * 3) * height;
case GST_VIDEO_FORMAT_Y41B:
- /* simplification of ROUNDUP4(w)*h + 2*((ROUNDUP8(w)/4)*h */
- return (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_8 (width) / 2)) * height;
+ /* simplification of ROUNDUP4(w)*h + 2*((ROUNDUP16(w)/4)*h */
+ return (GST_ROUND_UP_4 (width) + (GST_ROUND_UP_16 (width) / 2)) * height;
case GST_VIDEO_FORMAT_Y42B:
- /* simplification of ROUNDUP4(w)*h + 2*(ROUNDUP8(w)/2)*h: */
+ /* simplification of ROUNDUP4(w)*h + 2*(ROUNDUP8(w)/2)*h */
return (GST_ROUND_UP_4 (width) + GST_ROUND_UP_8 (width)) * height;
+ case GST_VIDEO_FORMAT_Y444:
+ return GST_ROUND_UP_4 (width) * height * 3;
+ case GST_VIDEO_FORMAT_v210:
+ return ((width + 47) / 48) * 128 * height;
+ case GST_VIDEO_FORMAT_v216:
+ return GST_ROUND_UP_8 (width * 4) * height;
+ case GST_VIDEO_FORMAT_NV12:
+ case GST_VIDEO_FORMAT_NV21:
+ return GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height) * 3 / 2;
+ case GST_VIDEO_FORMAT_GRAY8:
+ case GST_VIDEO_FORMAT_Y800:
+ case GST_VIDEO_FORMAT_RGB8_PALETTED:
+ return GST_ROUND_UP_4 (width) * height;
+ case GST_VIDEO_FORMAT_GRAY16_BE:
+ case GST_VIDEO_FORMAT_GRAY16_LE:
+ case GST_VIDEO_FORMAT_Y16:
+ return GST_ROUND_UP_4 (width * 2) * height;
+ case GST_VIDEO_FORMAT_UYVP:
+ return GST_ROUND_UP_4 ((width * 2 * 5 + 3) / 4) * height;
+ case GST_VIDEO_FORMAT_A420:
+ size = 2 * GST_ROUND_UP_4 (width) * GST_ROUND_UP_2 (height);
+ size += GST_ROUND_UP_4 (GST_ROUND_UP_2 (width) / 2) *
+ (GST_ROUND_UP_2 (height) / 2) * 2;
+ return size;
+ case GST_VIDEO_FORMAT_YUV9:
+ case GST_VIDEO_FORMAT_YVU9:
+ size = GST_ROUND_UP_4 (width) * height;
+ size += GST_ROUND_UP_4 (GST_ROUND_UP_4 (width) / 4) *
+ (GST_ROUND_UP_4 (height) / 4) * 2;
+ return size;
+ case GST_VIDEO_FORMAT_ARGB64:
+ case GST_VIDEO_FORMAT_AYUV64:
+ return width * 8 * height;
default:
return 0;
}
/**
* gst_video_format_convert:
- * @format: a #GstVideoFormat
- * @width: the width of video
- * @height: the height of video
- * @fps_n: frame rate numerator
- * @fps_d: frame rate denominator
+ * @info: a #GstVideoInfo
* @src_format: #GstFormat of the @src_value
* @src_value: value to convert
* @dest_format: #GstFormat of the @dest_value
* Converts among various #GstFormat types. This function handles
* GST_FORMAT_BYTES, GST_FORMAT_TIME, and GST_FORMAT_DEFAULT. For
* raw video, GST_FORMAT_DEFAULT corresponds to video frames. This
- * function can be to handle pad queries of the type GST_QUERY_CONVERT.
+ * function can be used to handle pad queries of the type GST_QUERY_CONVERT.
*
* Since: 0.10.16
*
* Returns: TRUE if the conversion was successful.
*/
gboolean
-gst_video_format_convert (GstVideoFormat format, int width, int height,
- int fps_n, int fps_d,
+gst_video_info_convert (GstVideoInfo * info,
GstFormat src_format, gint64 src_value,
GstFormat dest_format, gint64 * dest_value)
{
gboolean ret = FALSE;
- int size;
+ int size, fps_n, fps_d;
- g_return_val_if_fail (format != GST_VIDEO_FORMAT_UNKNOWN, 0);
- g_return_val_if_fail (width > 0 && height > 0, 0);
+ g_return_val_if_fail (info != NULL, 0);
+ g_return_val_if_fail (info->format != GST_VIDEO_FORMAT_UNKNOWN, 0);
+ g_return_val_if_fail (info->size > 0, 0);
- size = gst_video_format_get_size (format, width, height);
+ size = info->size;
+ fps_n = info->fps_n;
+ fps_d = info->fps_d;
GST_DEBUG ("converting value %" G_GINT64_FORMAT " from %s to %s",
src_value, gst_format_get_name (src_format),
return ret;
}
+
+#define GST_VIDEO_EVENT_STILL_STATE_NAME "GstEventStillFrame"
+
+/**
+ * gst_video_event_new_still_frame:
+ * @in_still: boolean value for the still-frame state of the event.
+ *
+ * Creates a new Still Frame event. If @in_still is %TRUE, then the event
+ * represents the start of a still frame sequence. If it is %FALSE, then
+ * the event ends a still frame sequence.
+ *
+ * To parse an event created by gst_video_event_new_still_frame() use
+ * gst_video_event_parse_still_frame().
+ *
+ * Returns: The new GstEvent
+ * Since: 0.10.26
+ */
+GstEvent *
+gst_video_event_new_still_frame (gboolean in_still)
+{
+ GstEvent *still_event;
+ GstStructure *s;
+
+ s = gst_structure_new (GST_VIDEO_EVENT_STILL_STATE_NAME,
+ "still-state", G_TYPE_BOOLEAN, in_still, NULL);
+ still_event = gst_event_new_custom (GST_EVENT_CUSTOM_DOWNSTREAM, s);
+
+ return still_event;
+}
+
+/**
+ * gst_video_event_parse_still_frame:
+ * @event: A #GstEvent to parse
+ * @in_still: A boolean to receive the still-frame status from the event, or NULL
+ *
+ * Parse a #GstEvent, identify if it is a Still Frame event, and
+ * return the still-frame state from the event if it is.
+ * If the event represents the start of a still frame, the in_still
+ * variable will be set to TRUE, otherwise FALSE. It is OK to pass NULL for the
+ * in_still variable order to just check whether the event is a valid still-frame
+ * event.
+ *
+ * Create a still frame event using gst_video_event_new_still_frame()
+ *
+ * Returns: %TRUE if the event is a valid still-frame event. %FALSE if not
+ * Since: 0.10.26
+ */
+gboolean
+gst_video_event_parse_still_frame (GstEvent * event, gboolean * in_still)
+{
+ const GstStructure *s;
+ gboolean ev_still_state;
+
+ g_return_val_if_fail (event != NULL, FALSE);
+
+ if (GST_EVENT_TYPE (event) != GST_EVENT_CUSTOM_DOWNSTREAM)
+ return FALSE; /* Not a still frame event */
+
+ s = gst_event_get_structure (event);
+ if (s == NULL
+ || !gst_structure_has_name (s, GST_VIDEO_EVENT_STILL_STATE_NAME))
+ return FALSE; /* Not a still frame event */
+ if (!gst_structure_get_boolean (s, "still-state", &ev_still_state))
+ return FALSE; /* Not a still frame event */
+ if (in_still)
+ *in_still = ev_still_state;
+ return TRUE;
+}
+
+#if 0
+/**
+ * gst_video_parse_caps_palette:
+ * @caps: #GstCaps to parse
+ *
+ * Returns the palette data from the caps as a #GstBuffer. For
+ * #GST_VIDEO_FORMAT_RGB8_PALETTED this is containing 256 #guint32
+ * values, each containing ARGB colors in native endianness.
+ *
+ * Returns: a #GstBuffer containing the palette data. Unref after usage.
+ * Since: 0.10.32
+ */
+GstBuffer *
+gst_video_parse_caps_palette (GstCaps * caps)
+{
+ GstStructure *s;
+ const GValue *p_v;
+ GstBuffer *p;
+
+ if (!gst_caps_is_fixed (caps))
+ return NULL;
+
+ s = gst_caps_get_structure (caps, 0);
+
+ p_v = gst_structure_get_value (s, "palette_data");
+ if (!p_v || !GST_VALUE_HOLDS_BUFFER (p_v))
+ return NULL;
+
+ p = gst_buffer_ref (gst_value_get_buffer (p_v));
+
+ return p;
+}
+#endif