v4l2codecs: Add v4l2 stateless H265 decoder
authorBenjamin Gaignard <benjamin.gaignard@collabora.com>
Tue, 12 Jul 2022 12:56:07 +0000 (08:56 -0400)
committerNicolas Dufresne <nicolas.dufresne@collabora.com>
Fri, 22 Jul 2022 14:21:02 +0000 (10:21 -0400)
Add v4l2 stateless H265 decoder.

Co-authored-by: Nicolas Dufresne <nicolas.dufresne@collabora.com>
Co-authored-by: Sebastian Fricke <sebastian.fricke@collabora.com>
Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/1079>

subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.c [new file with mode: 0644]
subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.h [new file with mode: 0644]
subprojects/gst-plugins-bad/sys/v4l2codecs/meson.build
subprojects/gst-plugins-bad/sys/v4l2codecs/plugin.c

diff --git a/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.c b/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.c
new file mode 100644 (file)
index 0000000..4594372
--- /dev/null
@@ -0,0 +1,1622 @@
+/* GStreamer
+ * Copyright (C) 2020 Nicolas Dufresne <nicolas.dufresne@collabora.com>
+ * Copyright (C) 2020 Safran Passenger Innovations LLC
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
+ * Boston, MA 02110-1301, USA.
+ */
+
+#ifdef HAVE_CONFIG_H
+#include <config.h>
+#endif
+
+#include "gstv4l2codecallocator.h"
+#include "gstv4l2codech265dec.h"
+#include "gstv4l2codecpool.h"
+#include "gstv4l2format.h"
+#include "linux/v4l2-controls.h"
+
+#define KERNEL_VERSION(a,b,c) (((a) << 16) + ((b) << 8) + (c))
+
+#define V4L2_MIN_KERNEL_VER_MAJOR 5
+#define V4L2_MIN_KERNEL_VER_MINOR 18
+#define V4L2_MIN_KERNEL_VERSION KERNEL_VERSION(V4L2_MIN_KERNEL_VER_MAJOR, V4L2_MIN_KERNEL_VER_MINOR, 0)
+
+GST_DEBUG_CATEGORY_STATIC (v4l2_h265dec_debug);
+#define GST_CAT_DEFAULT v4l2_h265dec_debug
+
+enum
+{
+  PROP_0,
+  PROP_LAST = PROP_0
+};
+
+static GstStaticPadTemplate sink_template =
+GST_STATIC_PAD_TEMPLATE (GST_VIDEO_DECODER_SINK_NAME,
+    GST_PAD_SINK, GST_PAD_ALWAYS,
+    GST_STATIC_CAPS ("video/x-h265, "
+        "stream-format=(string) { hvc1, hev1, byte-stream }, "
+        "alignment=(string) au")
+    );
+
+static GstStaticPadTemplate src_template =
+GST_STATIC_PAD_TEMPLATE (GST_VIDEO_DECODER_SRC_NAME,
+    GST_PAD_SRC, GST_PAD_ALWAYS,
+    GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (GST_V4L2_DEFAULT_VIDEO_FORMATS)));
+
+struct _GstV4l2CodecH265Dec
+{
+  GstH265Decoder parent;
+  GstV4l2Decoder *decoder;
+  GstVideoCodecState *output_state;
+  GstVideoInfo vinfo;
+  gint display_width;
+  gint display_height;
+  gint coded_width;
+  gint coded_height;
+  guint bitdepth;
+  guint chroma_format_idc;
+  guint num_slices;
+  gboolean first_slice;
+
+  GstV4l2CodecAllocator *sink_allocator;
+  GstV4l2CodecAllocator *src_allocator;
+  GstV4l2CodecPool *src_pool;
+  gint min_pool_size;
+  gboolean has_videometa;
+  gboolean need_negotiation;
+  gboolean copy_frames;
+  gboolean need_sequence;
+
+  struct v4l2_ctrl_hevc_sps sps;
+  struct v4l2_ctrl_hevc_pps pps;
+  struct v4l2_ctrl_hevc_scaling_matrix scaling_matrix;
+  struct v4l2_ctrl_hevc_decode_params decode_params;
+  GArray *slice_params;
+  GArray *entry_point_offsets;
+
+  enum v4l2_stateless_hevc_decode_mode decode_mode;
+  enum v4l2_stateless_hevc_start_code start_code;
+
+  GstMemory *bitstream;
+  GstMapInfo bitstream_map;
+
+  gboolean support_scaling_matrix;
+  gboolean support_slice_parameters;
+  gboolean support_entry_point_offsets;
+
+  GstVideoConverter *convert;
+  gboolean need_crop;
+  gint crop_rect_width, crop_rect_height;
+  gint crop_rect_x, crop_rect_y;
+};
+
+G_DEFINE_ABSTRACT_TYPE (GstV4l2CodecH265Dec, gst_v4l2_codec_h265_dec,
+    GST_TYPE_H265_DECODER);
+
+#define parent_class gst_v4l2_codec_h265_dec_parent_class
+
+static gboolean
+is_frame_based (GstV4l2CodecH265Dec * self)
+{
+  return (self->decode_mode ==
+      V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED) &&
+      !self->support_slice_parameters;
+}
+
+static gboolean
+is_slice_based (GstV4l2CodecH265Dec * self)
+{
+  return self->decode_mode == V4L2_STATELESS_HEVC_DECODE_MODE_SLICE_BASED;
+}
+
+static gboolean
+is_frame_based_with_slices (GstV4l2CodecH265Dec * self)
+{
+  return (self->decode_mode ==
+      V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED) &&
+      self->support_slice_parameters;
+}
+
+static gboolean
+needs_start_codes (GstV4l2CodecH265Dec * self)
+{
+  return self->start_code == V4L2_STATELESS_HEVC_START_CODE_ANNEX_B;
+}
+
+
+static gboolean
+gst_v4l2_codec_h265_dec_open (GstVideoDecoder * decoder)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  /* *INDENT-OFF* */
+  struct v4l2_ext_control control[] = {
+    {
+      .id = V4L2_CID_STATELESS_HEVC_DECODE_MODE,
+    },
+    {
+      .id = V4L2_CID_STATELESS_HEVC_START_CODE,
+    },
+  };
+  struct v4l2_ext_control scaling_matrix[] = {
+    {
+      .id = V4L2_CID_STATELESS_HEVC_SCALING_MATRIX,
+      .ptr = &self->scaling_matrix,
+      .size = sizeof (self->scaling_matrix),
+    },
+  };
+  /* *INDENT-ON* */
+
+  if (!gst_v4l2_decoder_open (self->decoder)) {
+    GST_ELEMENT_ERROR (self, RESOURCE, OPEN_READ_WRITE,
+        ("Failed to open H265 decoder"),
+        ("gst_v4l2_decoder_open() failed: %s", g_strerror (errno)));
+    return FALSE;
+  }
+
+  if (!gst_v4l2_decoder_get_controls (self->decoder, control,
+          G_N_ELEMENTS (control))) {
+    GST_ELEMENT_ERROR (self, RESOURCE, OPEN_READ_WRITE,
+        ("Driver did not report framing and start code method."),
+        ("gst_v4l2_decoder_get_controls() failed: %s", g_strerror (errno)));
+    return FALSE;
+  }
+
+  self->support_scaling_matrix =
+      gst_v4l2_decoder_get_controls (self->decoder, scaling_matrix,
+      G_N_ELEMENTS (scaling_matrix));
+
+  self->support_slice_parameters =
+      gst_v4l2_decoder_query_control_size (self->decoder,
+      V4L2_CID_STATELESS_HEVC_SLICE_PARAMS, NULL);
+
+  self->support_entry_point_offsets =
+      gst_v4l2_decoder_query_control_size (self->decoder,
+      V4L2_CID_STATELESS_HEVC_ENTRY_POINT_OFFSETS, NULL);
+
+  self->decode_mode = control[0].value;
+  self->start_code = control[1].value;
+
+  GST_INFO_OBJECT (self, "Opened H265 %s decoder %s",
+      is_frame_based (self) ? "frame based" : is_slice_based (self) ?
+      "slice based" : "frame based with slices",
+      needs_start_codes (self) ? "using start-codes" : "without start-codes");
+  gst_h265_decoder_set_process_ref_pic_lists (GST_H265_DECODER (self),
+      is_slice_based (self) || is_frame_based_with_slices (self));
+
+  return TRUE;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_close (GstVideoDecoder * decoder)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  gst_v4l2_decoder_close (self->decoder);
+  return TRUE;
+}
+
+static void
+gst_v4l2_codec_h265_dec_reset_allocation (GstV4l2CodecH265Dec * self)
+{
+  if (self->sink_allocator) {
+    gst_v4l2_codec_allocator_detach (self->sink_allocator);
+    g_clear_object (&self->sink_allocator);
+  }
+
+  if (self->src_allocator) {
+    gst_v4l2_codec_allocator_detach (self->src_allocator);
+    g_clear_object (&self->src_allocator);
+    g_clear_object (&self->src_pool);
+  }
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_stop (GstVideoDecoder * decoder)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+
+  gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SINK);
+  gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SRC);
+
+  gst_v4l2_codec_h265_dec_reset_allocation (self);
+
+  if (self->output_state)
+    gst_video_codec_state_unref (self->output_state);
+  self->output_state = NULL;
+
+  return GST_VIDEO_DECODER_CLASS (parent_class)->stop (decoder);
+}
+
+static gint
+get_pixel_bitdepth (GstV4l2CodecH265Dec * self)
+{
+  gint depth;
+
+  switch (self->chroma_format_idc) {
+    case 0:
+      /* 4:0:0 */
+      depth = self->bitdepth;
+      break;
+    case 1:
+      /* 4:2:0 */
+      depth = self->bitdepth + self->bitdepth / 2;
+      break;
+    case 2:
+      /* 4:2:2 */
+      depth = 2 * self->bitdepth;
+      break;
+    case 3:
+      /* 4:4:4 */
+      depth = 3 * self->bitdepth;
+      break;
+    default:
+      GST_WARNING_OBJECT (self, "Unsupported chroma format %i",
+          self->chroma_format_idc);
+      depth = 0;
+      break;
+  }
+
+  return depth;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_negotiate (GstVideoDecoder * decoder)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  GstH265Decoder *h265dec = GST_H265_DECODER (decoder);
+  /* *INDENT-OFF* */
+  struct v4l2_ext_control control[] = {
+    {
+      .id = V4L2_CID_STATELESS_HEVC_SPS,
+      .ptr = &self->sps,
+      .size = sizeof (self->sps),
+    },
+  };
+  /* *INDENT-ON* */
+  GstCaps *filter, *caps;
+
+  /* Ignore downstream renegotiation request. */
+  if (!self->need_negotiation)
+    return TRUE;
+  self->need_negotiation = FALSE;
+
+  GST_DEBUG_OBJECT (self, "Negotiate");
+
+  gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SINK);
+  gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SRC);
+
+  gst_v4l2_codec_h265_dec_reset_allocation (self);
+
+  if (!gst_v4l2_decoder_set_sink_fmt (self->decoder, V4L2_PIX_FMT_HEVC_SLICE,
+          self->coded_width, self->coded_height, get_pixel_bitdepth (self))) {
+    GST_ELEMENT_ERROR (self, CORE, NEGOTIATION,
+        ("Failed to configure H265 decoder"),
+        ("gst_v4l2_decoder_set_sink_fmt() failed: %s", g_strerror (errno)));
+    gst_v4l2_decoder_close (self->decoder);
+    return FALSE;
+  }
+
+  if (!gst_v4l2_decoder_set_controls (self->decoder, NULL, control,
+          G_N_ELEMENTS (control))) {
+    GST_ELEMENT_ERROR (decoder, RESOURCE, WRITE,
+        ("Driver does not support the selected stream."), (NULL));
+    return FALSE;
+  }
+
+  filter = gst_v4l2_decoder_enum_src_formats (self->decoder);
+  if (!filter) {
+    GST_ELEMENT_ERROR (self, CORE, NEGOTIATION,
+        ("No supported decoder output formats"), (NULL));
+    return FALSE;
+  }
+  GST_DEBUG_OBJECT (self, "Supported output formats: %" GST_PTR_FORMAT, filter);
+
+  caps = gst_pad_peer_query_caps (decoder->srcpad, filter);
+  gst_caps_unref (filter);
+  GST_DEBUG_OBJECT (self, "Peer supported formats: %" GST_PTR_FORMAT, caps);
+
+  if (!gst_v4l2_decoder_select_src_format (self->decoder, caps, &self->vinfo)) {
+    GST_ELEMENT_ERROR (self, CORE, NEGOTIATION,
+        ("Unsupported bitdepth/chroma format"),
+        ("No support for %ux%u %ubit chroma IDC %i", self->coded_width,
+            self->coded_height, self->bitdepth, self->chroma_format_idc));
+    gst_caps_unref (caps);
+    return FALSE;
+  }
+  gst_caps_unref (caps);
+
+  if (self->output_state)
+    gst_video_codec_state_unref (self->output_state);
+
+  self->output_state =
+      gst_video_decoder_set_output_state (GST_VIDEO_DECODER (self),
+      self->vinfo.finfo->format, self->display_width,
+      self->display_height, h265dec->input_state);
+
+  self->output_state->caps = gst_video_info_to_caps (&self->output_state->info);
+
+  if (GST_VIDEO_DECODER_CLASS (parent_class)->negotiate (decoder)) {
+    if (!gst_v4l2_decoder_streamon (self->decoder, GST_PAD_SINK)) {
+      GST_ELEMENT_ERROR (self, RESOURCE, FAILED,
+          ("Could not enable the decoder driver."),
+          ("VIDIOC_STREAMON(SINK) failed: %s", g_strerror (errno)));
+      return FALSE;
+    }
+
+    if (!gst_v4l2_decoder_streamon (self->decoder, GST_PAD_SRC)) {
+      GST_ELEMENT_ERROR (self, RESOURCE, FAILED,
+          ("Could not enable the decoder driver."),
+          ("VIDIOC_STREAMON(SRC) failed: %s", g_strerror (errno)));
+      return FALSE;
+    }
+
+    return TRUE;
+  }
+
+  return FALSE;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_decide_allocation (GstVideoDecoder * decoder,
+    GstQuery * query)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  guint min = 0;
+
+  self->has_videometa = gst_query_find_allocation_meta (query,
+      GST_VIDEO_META_API_TYPE, NULL);
+
+  g_clear_object (&self->src_pool);
+  g_clear_object (&self->src_allocator);
+
+  if (gst_query_get_n_allocation_pools (query) > 0)
+    gst_query_parse_nth_allocation_pool (query, 0, NULL, NULL, &min, NULL);
+
+  min = MAX (2, min);
+
+  self->sink_allocator = gst_v4l2_codec_allocator_new (self->decoder,
+      GST_PAD_SINK, self->min_pool_size + 2);
+  self->src_allocator = gst_v4l2_codec_allocator_new (self->decoder,
+      GST_PAD_SRC, self->min_pool_size + min + 1);
+  self->src_pool = gst_v4l2_codec_pool_new (self->src_allocator, &self->vinfo);
+
+  /* Our buffer pool is internal, we will let the base class create a video
+   * pool, and use it if we are running out of buffers or if downstream does
+   * not support GstVideoMeta */
+  return GST_VIDEO_DECODER_CLASS (parent_class)->decide_allocation
+      (decoder, query);
+}
+
+static void
+gst_v4l2_codec_h265_dec_fill_sequence (GstV4l2CodecH265Dec * self,
+    const GstH265SPS * sps)
+{
+  /* Whenever we update teh sps, we need to send it again */
+  self->need_sequence = TRUE;
+
+  /* *INDENT-OFF* */
+  self->sps = (struct v4l2_ctrl_hevc_sps) {
+        .video_parameter_set_id = sps->vps->id,
+        .seq_parameter_set_id = sps->id,
+       .pic_width_in_luma_samples = sps->pic_width_in_luma_samples,
+       .pic_height_in_luma_samples = sps->pic_height_in_luma_samples,
+       .bit_depth_luma_minus8 = sps->bit_depth_luma_minus8,
+       .bit_depth_chroma_minus8 = sps->bit_depth_chroma_minus8,
+       .log2_max_pic_order_cnt_lsb_minus4 = sps->log2_max_pic_order_cnt_lsb_minus4,
+       .sps_max_dec_pic_buffering_minus1 = sps->max_dec_pic_buffering_minus1[0],
+       .sps_max_num_reorder_pics = sps->max_num_reorder_pics[0],
+       .sps_max_latency_increase_plus1 = sps->max_latency_increase_plus1[0],
+       .log2_min_luma_coding_block_size_minus3 = sps->log2_min_luma_coding_block_size_minus3,
+       .log2_diff_max_min_luma_coding_block_size = sps->log2_diff_max_min_luma_coding_block_size,
+       .log2_min_luma_transform_block_size_minus2 = sps->log2_min_transform_block_size_minus2,
+       .log2_diff_max_min_luma_transform_block_size = sps->log2_diff_max_min_transform_block_size,
+       .max_transform_hierarchy_depth_inter = sps->max_transform_hierarchy_depth_inter,
+       .max_transform_hierarchy_depth_intra = sps->max_transform_hierarchy_depth_intra,
+       .num_short_term_ref_pic_sets = sps->num_short_term_ref_pic_sets,
+       .num_long_term_ref_pics_sps = sps->num_long_term_ref_pics_sps,
+        .chroma_format_idc = sps->chroma_format_idc,
+        .sps_max_sub_layers_minus1 = sps->max_sub_layers_minus1,
+        .flags = (sps->separate_colour_plane_flag ? V4L2_HEVC_SPS_FLAG_SEPARATE_COLOUR_PLANE : 0) |
+            (sps->scaling_list_enabled_flag ? V4L2_HEVC_SPS_FLAG_SCALING_LIST_ENABLED : 0) |
+            (sps->amp_enabled_flag ? V4L2_HEVC_SPS_FLAG_AMP_ENABLED : 0) |
+            (sps->sample_adaptive_offset_enabled_flag ? V4L2_HEVC_SPS_FLAG_SAMPLE_ADAPTIVE_OFFSET : 0) |
+            (sps->long_term_ref_pics_present_flag ? V4L2_HEVC_SPS_FLAG_LONG_TERM_REF_PICS_PRESENT : 0) |
+            (sps->temporal_mvp_enabled_flag ? V4L2_HEVC_SPS_FLAG_SPS_TEMPORAL_MVP_ENABLED : 0) |
+            (sps->strong_intra_smoothing_enabled_flag ? V4L2_HEVC_SPS_FLAG_STRONG_INTRA_SMOOTHING_ENABLED : 0),
+  };
+  /* *INDENT-ON* */
+  if (sps->pcm_enabled_flag) {
+    self->sps.pcm_sample_bit_depth_luma_minus1 =
+        sps->pcm_sample_bit_depth_luma_minus1;
+    self->sps.pcm_sample_bit_depth_chroma_minus1 =
+        sps->pcm_sample_bit_depth_chroma_minus1;
+    self->sps.log2_min_pcm_luma_coding_block_size_minus3 =
+        sps->log2_min_pcm_luma_coding_block_size_minus3;
+    self->sps.log2_diff_max_min_pcm_luma_coding_block_size =
+        sps->log2_diff_max_min_pcm_luma_coding_block_size;
+    self->sps.flags |=
+        V4L2_HEVC_SPS_FLAG_PCM_ENABLED | (sps->pcm_loop_filter_disabled_flag ?
+        V4L2_HEVC_SPS_FLAG_PCM_LOOP_FILTER_DISABLED : 0);
+  }
+}
+
+static void
+gst_v4l2_codec_h265_dec_fill_pps (GstV4l2CodecH265Dec * self, GstH265PPS * pps)
+{
+  gint i;
+
+  /* *INDENT-OFF* */
+  self->pps = (struct v4l2_ctrl_hevc_pps) {
+        .pic_parameter_set_id = pps->id,
+       .num_extra_slice_header_bits = pps->num_extra_slice_header_bits,
+       .num_ref_idx_l0_default_active_minus1 = pps->num_ref_idx_l0_default_active_minus1,
+       .num_ref_idx_l1_default_active_minus1 = pps->num_ref_idx_l1_default_active_minus1,
+       .init_qp_minus26 = pps->init_qp_minus26,
+       .diff_cu_qp_delta_depth = pps->diff_cu_qp_delta_depth,
+       .pps_cb_qp_offset = pps->cb_qp_offset,
+       .pps_cr_qp_offset = pps->cr_qp_offset,
+       .pps_beta_offset_div2 = pps->beta_offset_div2,
+       .pps_tc_offset_div2 = pps->tc_offset_div2,
+       .log2_parallel_merge_level_minus2 = pps->log2_parallel_merge_level_minus2,
+       .flags = (pps->dependent_slice_segments_enabled_flag ? V4L2_HEVC_PPS_FLAG_DEPENDENT_SLICE_SEGMENT_ENABLED : 0) |
+            (pps->output_flag_present_flag ? V4L2_HEVC_PPS_FLAG_OUTPUT_FLAG_PRESENT : 0) |
+            (pps->sign_data_hiding_enabled_flag ? V4L2_HEVC_PPS_FLAG_SIGN_DATA_HIDING_ENABLED : 0) |
+            (pps->cabac_init_present_flag ? V4L2_HEVC_PPS_FLAG_CABAC_INIT_PRESENT : 0) |
+            (pps->constrained_intra_pred_flag ? V4L2_HEVC_PPS_FLAG_CONSTRAINED_INTRA_PRED : 0) |
+            (pps->transform_skip_enabled_flag ? V4L2_HEVC_PPS_FLAG_TRANSFORM_SKIP_ENABLED : 0) |
+            (pps->cu_qp_delta_enabled_flag ? V4L2_HEVC_PPS_FLAG_CU_QP_DELTA_ENABLED : 0) |
+            (pps->slice_chroma_qp_offsets_present_flag ? V4L2_HEVC_PPS_FLAG_PPS_SLICE_CHROMA_QP_OFFSETS_PRESENT : 0) |
+            (pps->weighted_pred_flag ? V4L2_HEVC_PPS_FLAG_WEIGHTED_PRED : 0) |
+            (pps->weighted_bipred_flag ? V4L2_HEVC_PPS_FLAG_WEIGHTED_BIPRED : 0) |
+            (pps->transquant_bypass_enabled_flag ? V4L2_HEVC_PPS_FLAG_TRANSQUANT_BYPASS_ENABLED : 0) |
+            (pps->tiles_enabled_flag ? V4L2_HEVC_PPS_FLAG_TILES_ENABLED : 0) |
+            (pps->entropy_coding_sync_enabled_flag ? V4L2_HEVC_PPS_FLAG_ENTROPY_CODING_SYNC_ENABLED : 0) |
+            (pps->loop_filter_across_tiles_enabled_flag ? V4L2_HEVC_PPS_FLAG_LOOP_FILTER_ACROSS_TILES_ENABLED : 0) |
+            (pps->loop_filter_across_slices_enabled_flag ? V4L2_HEVC_PPS_FLAG_PPS_LOOP_FILTER_ACROSS_SLICES_ENABLED : 0 ) |
+            (pps->deblocking_filter_override_enabled_flag ? V4L2_HEVC_PPS_FLAG_DEBLOCKING_FILTER_OVERRIDE_ENABLED : 0) |
+            (pps->deblocking_filter_disabled_flag ? V4L2_HEVC_PPS_FLAG_PPS_DISABLE_DEBLOCKING_FILTER : 0) |
+            (pps->lists_modification_present_flag ? V4L2_HEVC_PPS_FLAG_LISTS_MODIFICATION_PRESENT : 0) |
+            (pps->slice_segment_header_extension_present_flag ? V4L2_HEVC_PPS_FLAG_SLICE_SEGMENT_HEADER_EXTENSION_PRESENT : 0) |
+            (pps->deblocking_filter_control_present_flag ? V4L2_HEVC_PPS_FLAG_DEBLOCKING_FILTER_CONTROL_PRESENT : 0) |
+            (pps->uniform_spacing_flag ? V4L2_HEVC_PPS_FLAG_UNIFORM_SPACING : 0),
+  };
+  /* *INDENT-ON* */
+
+  if (pps->tiles_enabled_flag) {
+    self->pps.num_tile_columns_minus1 = pps->num_tile_columns_minus1;
+    self->pps.num_tile_rows_minus1 = pps->num_tile_rows_minus1;
+
+    /* This should not be needed if we use uniform spacing, but Cedrus driver
+     * depends on it for now. As GStreamer parser do set that, lets just pass
+     * the information we have. */
+    for (i = 0; i <= pps->num_tile_columns_minus1; i++)
+      self->pps.column_width_minus1[i] = pps->column_width_minus1[i];
+    for (i = 0; i <= pps->num_tile_rows_minus1; i++)
+      self->pps.row_height_minus1[i] = pps->row_height_minus1[i];
+  }
+}
+
+static void
+gst_v4l2_codec_h265_dec_fill_scaling_matrix (GstV4l2CodecH265Dec * self,
+    GstH265PPS * pps)
+{
+  GstH265ScalingList *sl;
+  gint i;
+
+  if (!pps->sps->scaling_list_enabled_flag) {
+    memset (&self->scaling_matrix, 0, sizeof (self->scaling_matrix));
+    return;
+  }
+
+  if (pps->scaling_list_data_present_flag)
+    sl = &pps->scaling_list;
+  else if (pps->sps->scaling_list_data_present_flag)
+    sl = &pps->sps->scaling_list;
+  /* The default scaling list is strored in the pps */
+  else
+    sl = &pps->scaling_list;
+
+  for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_4x4); i++)
+    gst_h265_quant_matrix_4x4_get_raster_from_uprightdiagonal
+        (self->scaling_matrix.scaling_list_4x4[i], sl->scaling_lists_4x4[i]);
+
+  for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_8x8); i++)
+    gst_h265_quant_matrix_8x8_get_raster_from_uprightdiagonal
+        (self->scaling_matrix.scaling_list_8x8[i], sl->scaling_lists_8x8[i]);
+
+  for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_16x16); i++)
+    gst_h265_quant_matrix_16x16_get_raster_from_uprightdiagonal
+        (self->scaling_matrix.scaling_list_16x16[i],
+        sl->scaling_lists_16x16[i]);
+
+  for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_32x32); i++)
+    gst_h265_quant_matrix_32x32_get_raster_from_uprightdiagonal
+        (self->scaling_matrix.scaling_list_32x32[i],
+        sl->scaling_lists_32x32[i]);
+
+  for (i = 0; i < G_N_ELEMENTS (sl->scaling_list_dc_coef_minus8_16x16); i++)
+    self->scaling_matrix.scaling_list_dc_coef_16x16[i] =
+        sl->scaling_list_dc_coef_minus8_16x16[i] + 8;
+
+  for (i = 0; i < G_N_ELEMENTS (sl->scaling_list_dc_coef_minus8_32x32); i++)
+    self->scaling_matrix.scaling_list_dc_coef_32x32[i] =
+        sl->scaling_list_dc_coef_minus8_32x32[i] + 8;
+}
+
+static guint
+get_slice_header_byte_offset (GstH265Slice * slice)
+{
+  guint epb_count, nal_header_bytes;
+
+  epb_count = slice->header.n_emulation_prevention_bytes;
+  nal_header_bytes = slice->nalu.header_bytes;
+
+  return nal_header_bytes + (slice->header.header_size + 7) / 8 - epb_count;
+}
+
+static void
+gst_v4l2_codec_h265_dec_fill_slice_params (GstV4l2CodecH265Dec * self,
+    GstH265Slice * slice, GstH265Picture * picture)
+{
+  gint n = self->num_slices++;
+  GstH265SliceHdr *slice_hdr = &slice->header;
+  GstH265PPS *pps = slice_hdr->pps;
+  gsize slice_size = slice->nalu.size;
+  gsize sc_offset = 0;
+  struct v4l2_ctrl_hevc_slice_params *params;
+  gint i, j;
+  gint chroma_weight, chroma_log2_weight_denom;
+  /* TODO adjust this if sps_ext is later supported */
+  const gint32 WpOffsetHalfRangeC = 1 << 7;
+
+  /* Ensure array is large enough */
+  if (self->slice_params->len < self->num_slices)
+    g_array_set_size (self->slice_params, self->slice_params->len * 2);
+
+  if (needs_start_codes (self))
+    sc_offset = 3;
+
+  /* *INDENT-OFF* */
+  params = &g_array_index (self->slice_params, struct v4l2_ctrl_hevc_slice_params, n);
+  *params = (struct v4l2_ctrl_hevc_slice_params) {
+    .bit_size = (slice_size + sc_offset) * 8,
+    .data_byte_offset = get_slice_header_byte_offset (slice) + sc_offset,
+    .num_entry_point_offsets = slice_hdr->num_entry_point_offsets,
+    .nal_unit_type = slice->nalu.type,
+    .nuh_temporal_id_plus1 = slice->nalu.temporal_id_plus1,
+    .slice_type = slice_hdr->type,
+    .colour_plane_id = slice_hdr->colour_plane_id,
+    .slice_pic_order_cnt = picture->pic_order_cnt,
+    .num_ref_idx_l0_active_minus1 = slice_hdr->num_ref_idx_l0_active_minus1,
+    .num_ref_idx_l1_active_minus1 = slice_hdr->num_ref_idx_l1_active_minus1,
+    .collocated_ref_idx = slice_hdr->collocated_ref_idx,
+    .five_minus_max_num_merge_cand = slice_hdr->five_minus_max_num_merge_cand,
+    .slice_qp_delta = slice_hdr->qp_delta,
+    .slice_cb_qp_offset = slice_hdr->cb_qp_offset,
+    .slice_cr_qp_offset = slice_hdr->cr_qp_offset,
+    .slice_act_y_qp_offset = slice_hdr->slice_act_y_qp_offset,
+    .slice_act_cb_qp_offset = slice_hdr->slice_act_cb_qp_offset,
+    .slice_act_cr_qp_offset = slice_hdr->slice_act_cr_qp_offset,
+    .slice_beta_offset_div2 = slice_hdr->beta_offset_div2,
+    .slice_tc_offset_div2 = slice_hdr->tc_offset_div2,
+    .pic_struct = picture->pic_struct,
+    .slice_segment_addr = slice_hdr->segment_address,
+    .short_term_ref_pic_set_size = slice_hdr->short_term_ref_pic_set_size,
+    .long_term_ref_pic_set_size = slice_hdr->long_term_ref_pic_set_size,
+    .pred_weight_table = (struct v4l2_hevc_pred_weight_table) {
+        .luma_log2_weight_denom = slice_hdr->pred_weight_table.luma_log2_weight_denom,
+        .delta_chroma_log2_weight_denom = slice_hdr->pred_weight_table.delta_chroma_log2_weight_denom,
+    },
+    .flags =
+        (slice_hdr->sao_luma_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_SAO_LUMA : 0) |
+        (slice_hdr->sao_chroma_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_SAO_CHROMA : 0) |
+        (slice_hdr->temporal_mvp_enabled_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_TEMPORAL_MVP_ENABLED : 0) |
+        (slice_hdr->mvd_l1_zero_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_MVD_L1_ZERO : 0) |
+        (slice_hdr->cabac_init_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_CABAC_INIT : 0) |
+        (slice_hdr->collocated_from_l0_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_COLLOCATED_FROM_L0 : 0) |
+        (slice_hdr->use_integer_mv_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_USE_INTEGER_MV : 0) |
+        (slice_hdr->deblocking_filter_disabled_flag ?
+             V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_DEBLOCKING_FILTER_DISABLED : 0) |
+        (slice_hdr->loop_filter_across_slices_enabled_flag ?
+             V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_LOOP_FILTER_ACROSS_SLICES_ENABLED : 0) |
+        (slice_hdr->dependent_slice_segment_flag ?
+             V4L2_HEVC_SLICE_PARAMS_FLAG_DEPENDENT_SLICE_SEGMENT : 0),
+  };
+  /* *INDENT-ON* */
+
+  for (i = 0; i < slice_hdr->num_entry_point_offsets; i++) {
+    guint32 entry_point_offset = slice_hdr->entry_point_offset_minus1[i] + 1;
+    g_array_append_val (self->entry_point_offsets, entry_point_offset);
+  }
+
+  if (GST_H265_IS_I_SLICE (slice_hdr) ||
+      (!pps->weighted_pred_flag && GST_H265_IS_P_SLICE (slice_hdr)) ||
+      (!pps->weighted_bipred_flag && GST_H265_IS_B_SLICE (slice_hdr)))
+    return;
+
+  for (i = 0; i <= slice_hdr->num_ref_idx_l0_active_minus1; i++) {
+    if (!slice_hdr->pred_weight_table.luma_weight_l0_flag[i])
+      continue;
+
+    params->pred_weight_table.delta_luma_weight_l0[i] =
+        slice_hdr->pred_weight_table.delta_luma_weight_l0[i];
+    params->pred_weight_table.luma_offset_l0[i] =
+        slice_hdr->pred_weight_table.luma_offset_l0[i];
+  }
+
+  chroma_log2_weight_denom =
+      slice_hdr->pred_weight_table.luma_log2_weight_denom +
+      slice_hdr->pred_weight_table.delta_chroma_log2_weight_denom;
+
+  if (slice_hdr->pps->sps->chroma_array_type != 0) {
+
+    for (i = 0; i <= slice_hdr->num_ref_idx_l0_active_minus1; i++) {
+      if (!slice_hdr->pred_weight_table.chroma_weight_l0_flag[i])
+        continue;
+
+      for (j = 0; j < 2; j++) {
+        gint16 delta_chroma_offset_l0 =
+            slice_hdr->pred_weight_table.delta_chroma_offset_l0[i][j];
+        gint chroma_offset;
+
+        params->pred_weight_table.delta_chroma_weight_l0[i][j] =
+            slice_hdr->pred_weight_table.delta_chroma_weight_l0[i][j];
+
+        /* Find  ChromaWeightL0 */
+        chroma_weight = (1 << chroma_log2_weight_denom) +
+            slice_hdr->pred_weight_table.delta_chroma_weight_l0[i][j];
+        chroma_offset = WpOffsetHalfRangeC + delta_chroma_offset_l0 -
+            ((WpOffsetHalfRangeC * chroma_weight) >> chroma_log2_weight_denom);
+
+        /* 7-56 */
+        params->pred_weight_table.chroma_offset_l0[i][j] =
+            CLAMP (chroma_offset, -WpOffsetHalfRangeC, WpOffsetHalfRangeC - 1);
+      }
+    }
+  }
+
+  /* Skip l1 if this is not a B-Frame. */
+  if (!GST_H265_IS_B_SLICE (slice_hdr))
+    return;
+
+  for (i = 0; i <= slice_hdr->num_ref_idx_l1_active_minus1; i++) {
+    if (!slice_hdr->pred_weight_table.luma_weight_l1_flag[i])
+      continue;
+
+    params->pred_weight_table.delta_luma_weight_l1[i] =
+        slice_hdr->pred_weight_table.delta_luma_weight_l1[i];
+    params->pred_weight_table.luma_offset_l1[i] =
+        slice_hdr->pred_weight_table.luma_offset_l1[i];
+  }
+
+  if (slice_hdr->pps->sps->chroma_array_type != 0) {
+    for (i = 0; i <= slice_hdr->num_ref_idx_l1_active_minus1; i++) {
+      if (!slice_hdr->pred_weight_table.chroma_weight_l1_flag[i])
+        continue;
+
+      for (j = 0; j < 2; j++) {
+        gint16 delta_chroma_offset_l1 =
+            slice_hdr->pred_weight_table.delta_chroma_offset_l1[i][j];
+        gint chroma_offset;
+
+        params->pred_weight_table.delta_chroma_weight_l1[i][j] =
+            slice_hdr->pred_weight_table.delta_chroma_weight_l1[i][j];
+
+        /* Find  ChromaWeightL1 */
+        chroma_weight = (1 << chroma_log2_weight_denom) +
+            slice_hdr->pred_weight_table.delta_chroma_weight_l1[i][j];
+
+        chroma_offset = WpOffsetHalfRangeC + delta_chroma_offset_l1 -
+            ((WpOffsetHalfRangeC * chroma_weight) >> chroma_log2_weight_denom);
+
+        /* 7-56 */
+        params->pred_weight_table.chroma_offset_l1[i][j] =
+            CLAMP (chroma_offset, -WpOffsetHalfRangeC, WpOffsetHalfRangeC - 1);
+      }
+    }
+  }
+}
+
+static guint8
+lookup_dpb_index (struct v4l2_hevc_dpb_entry dpb[16], GstH265Picture * ref_pic)
+{
+  guint64 ref_ts;
+  gint i;
+
+  /* Reference list may have wholes in case a ref is missing, we should mark
+   * the whole and avoid moving items in the list */
+  if (!ref_pic)
+    return 0xff;
+
+  ref_ts = (guint64) ref_pic->system_frame_number * 1000;
+  for (i = 0; i < 16; i++) {
+    if (dpb[i].timestamp == ref_ts)
+      return i;
+  }
+
+  return 0xff;
+}
+
+static void
+gst_v4l2_codec_h265_dec_fill_references (GstV4l2CodecH265Dec * self,
+    GArray * ref_pic_list0, GArray * ref_pic_list1)
+{
+  struct v4l2_ctrl_hevc_slice_params *slice_params;
+  struct v4l2_ctrl_hevc_decode_params *decode_params = &self->decode_params;
+  gint i;
+
+  slice_params = &g_array_index (self->slice_params,
+      struct v4l2_ctrl_hevc_slice_params, self->num_slices - 1);
+
+  memset (slice_params->ref_idx_l0, 0xff, sizeof (slice_params->ref_idx_l0));
+  memset (slice_params->ref_idx_l1, 0xff, sizeof (slice_params->ref_idx_l1));
+
+  for (i = 0; i < ref_pic_list0->len; i++) {
+    GstH265Picture *ref_pic =
+        g_array_index (ref_pic_list0, GstH265Picture *, i);
+    slice_params->ref_idx_l0[i] =
+        lookup_dpb_index (decode_params->dpb, ref_pic);
+  }
+
+  for (i = 0; i < ref_pic_list1->len; i++) {
+    GstH265Picture *ref_pic =
+        g_array_index (ref_pic_list1, GstH265Picture *, i);
+    slice_params->ref_idx_l1[i] =
+        lookup_dpb_index (decode_params->dpb, ref_pic);
+  }
+}
+
+static GstFlowReturn
+gst_v4l2_codec_h265_dec_new_sequence (GstH265Decoder * decoder,
+    const GstH265SPS * sps, gint max_dpb_size)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  gint crop_width = sps->width;
+  gint crop_height = sps->height;
+  gboolean negotiation_needed = FALSE;
+
+  if (self->vinfo.finfo->format == GST_VIDEO_FORMAT_UNKNOWN)
+    negotiation_needed = TRUE;
+
+  /* TODO check if CREATE_BUFS is supported, and simply grow the pool */
+  if (self->min_pool_size < max_dpb_size) {
+    self->min_pool_size = max_dpb_size;
+    negotiation_needed = TRUE;
+  }
+
+  self->need_crop = FALSE;
+  if (sps->conformance_window_flag) {
+    crop_width = sps->crop_rect_width;
+    crop_height = sps->crop_rect_height;
+    self->crop_rect_width = sps->crop_rect_width;
+    self->crop_rect_height = sps->crop_rect_height;
+    self->crop_rect_x = sps->crop_rect_x;
+    self->crop_rect_y = sps->crop_rect_y;
+
+    /* conformance_window_flag could be set but with zeroed
+     * parameters so check if we really need to crop */
+    self->need_crop |= self->crop_rect_width != sps->width;
+    self->need_crop |= self->crop_rect_height != sps->height;
+    self->need_crop |= self->crop_rect_x != 0;
+    self->need_crop |= self->crop_rect_y != 0;
+  }
+
+  /* TODO Check if current buffers are large enough, and reuse them */
+  if (self->display_width != crop_width || self->display_height != crop_height
+      || self->coded_width != sps->width || self->coded_height != sps->height) {
+    self->display_width = crop_width;
+    self->display_height = crop_height;
+    self->coded_width = sps->width;
+    self->coded_height = sps->height;
+    negotiation_needed = TRUE;
+    GST_INFO_OBJECT (self, "Resolution changed to %dx%d (%ix%i)",
+        self->display_width, self->display_height,
+        self->coded_width, self->coded_height);
+  }
+
+  if (self->bitdepth != sps->bit_depth_luma_minus8 + 8) {
+    self->bitdepth = sps->bit_depth_luma_minus8 + 8;
+    negotiation_needed = TRUE;
+    GST_INFO_OBJECT (self, "Bitdepth changed to %u", self->bitdepth);
+  }
+
+  if (self->chroma_format_idc != sps->chroma_format_idc) {
+    self->chroma_format_idc = sps->chroma_format_idc;
+    negotiation_needed = TRUE;
+    GST_INFO_OBJECT (self, "Chroma format changed to %i",
+        self->chroma_format_idc);
+  }
+
+  gst_v4l2_codec_h265_dec_fill_sequence (self, sps);
+
+  if (negotiation_needed) {
+    self->need_negotiation = TRUE;
+    if (!gst_video_decoder_negotiate (GST_VIDEO_DECODER (self))) {
+      GST_ERROR_OBJECT (self, "Failed to negotiate with downstream");
+      return GST_FLOW_NOT_NEGOTIATED;
+    }
+  }
+
+  /* Check if we can zero-copy buffers */
+  if (!self->has_videometa) {
+    GstVideoInfo ref_vinfo;
+    gint i;
+
+    gst_video_info_set_format (&ref_vinfo, GST_VIDEO_INFO_FORMAT (&self->vinfo),
+        self->display_width, self->display_height);
+
+    for (i = 0; i < GST_VIDEO_INFO_N_PLANES (&self->vinfo); i++) {
+      if (self->vinfo.stride[i] != ref_vinfo.stride[i] ||
+          self->vinfo.offset[i] != ref_vinfo.offset[i]) {
+        GST_WARNING_OBJECT (self,
+            "GstVideoMeta support required, copying frames.");
+        self->copy_frames = TRUE;
+        break;
+      }
+    }
+  } else {
+    self->copy_frames = self->need_crop;
+  }
+
+  return GST_FLOW_OK;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_ensure_bitstream (GstV4l2CodecH265Dec * self)
+{
+  if (self->bitstream)
+    goto done;
+
+  self->bitstream = gst_v4l2_codec_allocator_alloc (self->sink_allocator);
+
+  if (!self->bitstream) {
+    GST_ELEMENT_ERROR (self, RESOURCE, NO_SPACE_LEFT,
+        ("Not enough memory to decode H265 stream."), (NULL));
+    return FALSE;
+  }
+
+  if (!gst_memory_map (self->bitstream, &self->bitstream_map, GST_MAP_WRITE)) {
+    GST_ELEMENT_ERROR (self, RESOURCE, WRITE,
+        ("Could not access bitstream memory for writing"), (NULL));
+    g_clear_pointer (&self->bitstream, gst_memory_unref);
+    return FALSE;
+  }
+
+done:
+  /* We use this field to track how much we have written */
+  self->bitstream_map.size = 0;
+
+  return TRUE;
+}
+
+static void
+gst_v4l2_codec_h265_dec_fill_decode_params (GstV4l2CodecH265Dec * self,
+    GstH265Picture * picture, GstH265Slice * slice, GstH265Dpb * dpb)
+{
+  GstH265Decoder *decoder = (GstH265Decoder *) self;
+  GArray *refs = gst_h265_dpb_get_pictures_all (dpb);
+  gint i;
+
+  /* *INDENT-OFF* */
+  self->decode_params = (struct v4l2_ctrl_hevc_decode_params) {
+    .pic_order_cnt_val = picture->pic_order_cnt,
+    .num_poc_st_curr_before = decoder->NumPocStCurrBefore,
+    .num_poc_st_curr_after = decoder->NumPocStCurrAfter,
+    .num_poc_lt_curr = decoder->NumPocLtCurr,
+    .flags =
+      (GST_H265_IS_NAL_TYPE_IRAP (slice->nalu.type) ? V4L2_HEVC_DECODE_PARAM_FLAG_IRAP_PIC : 0) |
+      (GST_H265_IS_NAL_TYPE_IDR (slice->nalu.type) ? V4L2_HEVC_DECODE_PARAM_FLAG_IDR_PIC : 0) |
+      (slice->header.no_output_of_prior_pics_flag ? V4L2_HEVC_DECODE_PARAM_FLAG_NO_OUTPUT_OF_PRIOR : 0),
+  };
+  /* *INDENT-ON* */
+
+  for (i = 0; i < refs->len; i++) {
+    GstH265Picture *ref_pic = g_array_index (refs, GstH265Picture *, i);
+
+    if (!ref_pic->ref)
+      continue;
+
+    /* *INDENT-OFF* */
+    self->decode_params.dpb[self->decode_params.num_active_dpb_entries++] =
+        (struct v4l2_hevc_dpb_entry) {
+      /*
+       * The reference is multiplied by 1000 because it's wassed as micro
+       * seconds and this TS is nanosecond.
+       */
+      .timestamp = (guint64) ref_pic->system_frame_number * 1000,
+      .flags = ref_pic->long_term ? V4L2_HEVC_DPB_ENTRY_LONG_TERM_REFERENCE : 0,
+      .field_pic = ref_pic->pic_struct,
+      .pic_order_cnt_val = ref_pic->pic_order_cnt,
+    };
+    /* *INDENT-ON* */
+  }
+
+  for (i = 0; i < 16; i++) {
+    self->decode_params.poc_st_curr_before[i] =
+        lookup_dpb_index (self->decode_params.dpb,
+        decoder->RefPicSetStCurrBefore[i]);
+    self->decode_params.poc_st_curr_after[i] =
+        lookup_dpb_index (self->decode_params.dpb,
+        decoder->RefPicSetStCurrAfter[i]);
+    self->decode_params.poc_lt_curr[i] =
+        lookup_dpb_index (self->decode_params.dpb, decoder->RefPicSetLtCurr[i]);
+  }
+
+  g_array_unref (refs);
+}
+
+static GstFlowReturn
+gst_v4l2_codec_h265_dec_start_picture (GstH265Decoder * decoder,
+    GstH265Picture * picture, GstH265Slice * slice, GstH265Dpb * dpb)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+
+  /* FIXME base class should not call us if negotiation failed */
+  if (!self->sink_allocator)
+    return GST_FLOW_NOT_NEGOTIATED;
+
+  if (!gst_v4l2_codec_h265_dec_ensure_bitstream (self))
+    return GST_FLOW_ERROR;
+
+  /* The base class will only emit new_sequence for allocation related changes
+   * in the SPS, make sure to keep the SPS upt-to-date */
+  if (slice->header.pps->sps->id != self->sps.seq_parameter_set_id)
+    gst_v4l2_codec_h265_dec_fill_sequence (self, slice->header.pps->sps);
+
+  gst_v4l2_codec_h265_dec_fill_pps (self, slice->header.pps);
+  gst_v4l2_codec_h265_dec_fill_scaling_matrix (self, slice->header.pps);
+  gst_v4l2_codec_h265_dec_fill_decode_params (self, picture, slice, dpb);
+
+  self->first_slice = TRUE;
+  self->num_slices = 0;
+  g_array_set_size (self->entry_point_offsets, 0);
+
+  return GST_FLOW_OK;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_crop_output_buffer (GstV4l2CodecH265Dec * self,
+    GstVideoFrame * dest_frame, GstVideoFrame * src_frame)
+{
+  GstVideoInfo dst_info = dest_frame->info;
+
+  dst_info.fps_n = src_frame->info.fps_n;
+  dst_info.fps_d = src_frame->info.fps_d;
+
+  if (self->convert) {
+    gboolean new_convert = FALSE;
+    gint x = 0, y = 0, width = 0, height = 0;
+    const GstStructure *config = gst_video_converter_get_config (self->convert);
+
+    if (!gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_X, &x)
+        || !gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_Y, &y)
+        || !gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_WIDTH,
+            &width)
+        || !gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_HEIGHT,
+            &height))
+      new_convert = TRUE;
+
+    new_convert |= (self->crop_rect_x != x);
+    new_convert |= (self->crop_rect_y != y);
+    new_convert |= (self->crop_rect_width != width);
+    new_convert |= (self->crop_rect_height != height);
+
+    /* No need to check dest, it always has (0,0) -> (width, height) */
+
+    if (new_convert)
+      g_clear_pointer (&self->convert, gst_video_converter_free);
+  }
+
+  if (!self->convert) {
+    self->convert = gst_video_converter_new (&src_frame->info, &dst_info,
+        gst_structure_new ("options",
+            GST_VIDEO_CONVERTER_OPT_DITHER_METHOD,
+            GST_TYPE_VIDEO_DITHER_METHOD, GST_VIDEO_DITHER_NONE,
+            GST_VIDEO_CONVERTER_OPT_DITHER_QUANTIZATION,
+            G_TYPE_UINT, 0,
+            GST_VIDEO_CONVERTER_OPT_CHROMA_MODE,
+            GST_TYPE_VIDEO_CHROMA_MODE, GST_VIDEO_CHROMA_MODE_NONE,
+            GST_VIDEO_CONVERTER_OPT_MATRIX_MODE,
+            GST_TYPE_VIDEO_MATRIX_MODE, GST_VIDEO_MATRIX_MODE_NONE,
+            GST_VIDEO_CONVERTER_OPT_SRC_X, G_TYPE_INT, self->crop_rect_x,
+            GST_VIDEO_CONVERTER_OPT_SRC_Y, G_TYPE_INT, self->crop_rect_y,
+            GST_VIDEO_CONVERTER_OPT_SRC_WIDTH, G_TYPE_INT,
+            self->crop_rect_width, GST_VIDEO_CONVERTER_OPT_SRC_HEIGHT,
+            G_TYPE_INT, self->crop_rect_height, GST_VIDEO_CONVERTER_OPT_DEST_X,
+            G_TYPE_INT, 0, GST_VIDEO_CONVERTER_OPT_DEST_Y, G_TYPE_INT, 0,
+            GST_VIDEO_CONVERTER_OPT_DEST_WIDTH, G_TYPE_INT, self->display_width,
+            GST_VIDEO_CONVERTER_OPT_DEST_HEIGHT, G_TYPE_INT,
+            self->display_height, NULL));
+
+    if (!self->convert) {
+      GST_WARNING_OBJECT (self, "failed to create a video convert");
+      return FALSE;
+    }
+  }
+
+  gst_video_converter_frame (self->convert, src_frame, dest_frame);
+
+  return TRUE;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_copy_output_buffer (GstV4l2CodecH265Dec * self,
+    GstVideoCodecFrame * codec_frame)
+{
+  GstVideoFrame src_frame;
+  GstVideoFrame dest_frame;
+  GstVideoInfo dest_vinfo;
+  GstBuffer *buffer;
+
+  gst_video_info_set_format (&dest_vinfo, GST_VIDEO_INFO_FORMAT (&self->vinfo),
+      self->display_width, self->display_height);
+
+  buffer = gst_video_decoder_allocate_output_buffer (GST_VIDEO_DECODER (self));
+  if (!buffer)
+    goto fail;
+
+  if (!gst_video_frame_map (&src_frame, &self->vinfo,
+          codec_frame->output_buffer, GST_MAP_READ))
+    goto fail;
+
+  if (!gst_video_frame_map (&dest_frame, &dest_vinfo, buffer, GST_MAP_WRITE)) {
+    gst_video_frame_unmap (&dest_frame);
+    goto fail;
+  }
+
+  if (self->need_crop) {
+    if (!gst_v4l2_codec_h265_dec_crop_output_buffer (self, &dest_frame,
+            &src_frame)) {
+      gst_video_frame_unmap (&src_frame);
+      gst_video_frame_unmap (&dest_frame);
+      GST_ERROR_OBJECT (self, "fail to apply the video crop.");
+      goto fail;
+    }
+  } else {
+    /* gst_video_frame_copy can crop this, but does not know, so let make it
+     * think it's all right */
+    GST_VIDEO_INFO_WIDTH (&src_frame.info) = self->display_width;
+    GST_VIDEO_INFO_HEIGHT (&src_frame.info) = self->display_height;
+
+    if (!gst_video_frame_copy (&dest_frame, &src_frame)) {
+      gst_video_frame_unmap (&src_frame);
+      gst_video_frame_unmap (&dest_frame);
+      goto fail;
+    }
+  }
+
+  gst_video_frame_unmap (&src_frame);
+  gst_video_frame_unmap (&dest_frame);
+  gst_buffer_replace (&codec_frame->output_buffer, buffer);
+  gst_buffer_unref (buffer);
+
+  return TRUE;
+
+fail:
+  GST_ERROR_OBJECT (self, "Failed copy output buffer.");
+  return FALSE;
+}
+
+static GstFlowReturn
+gst_v4l2_codec_h265_dec_output_picture (GstH265Decoder * decoder,
+    GstVideoCodecFrame * frame, GstH265Picture * picture)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  GstV4l2Request *request = gst_h265_picture_get_user_data (picture);
+  gint ret;
+
+  GST_DEBUG_OBJECT (self, "Output picture %u", picture->system_frame_number);
+
+  ret = gst_v4l2_request_set_done (request);
+  if (ret == 0) {
+    GST_ELEMENT_ERROR (self, STREAM, DECODE,
+        ("Decoding frame %u took too long", picture->system_frame_number),
+        (NULL));
+    goto error;
+  } else if (ret < 0) {
+    GST_ELEMENT_ERROR (self, STREAM, DECODE,
+        ("Decoding request failed: %s", g_strerror (errno)), (NULL));
+    goto error;
+  }
+  g_return_val_if_fail (frame->output_buffer, GST_FLOW_ERROR);
+
+  if (gst_v4l2_request_failed (request)) {
+    GST_ELEMENT_ERROR (self, STREAM, DECODE,
+        ("Failed to decode frame %u", picture->system_frame_number), (NULL));
+    goto error;
+  }
+
+  /* Hold on reference buffers for the rest of the picture lifetime */
+  gst_h265_picture_set_user_data (picture,
+      gst_buffer_ref (frame->output_buffer), (GDestroyNotify) gst_buffer_unref);
+
+  if (self->copy_frames)
+    gst_v4l2_codec_h265_dec_copy_output_buffer (self, frame);
+
+  gst_h265_picture_unref (picture);
+
+  return gst_video_decoder_finish_frame (GST_VIDEO_DECODER (self), frame);
+
+error:
+  gst_video_decoder_drop_frame (GST_VIDEO_DECODER (self), frame);
+  gst_h265_picture_unref (picture);
+
+  return GST_FLOW_ERROR;
+}
+
+static void
+gst_v4l2_codec_h265_dec_reset_picture (GstV4l2CodecH265Dec * self)
+{
+  if (self->bitstream) {
+    if (self->bitstream_map.memory)
+      gst_memory_unmap (self->bitstream, &self->bitstream_map);
+    g_clear_pointer (&self->bitstream, gst_memory_unref);
+    self->bitstream_map = (GstMapInfo) GST_MAP_INFO_INIT;
+  }
+
+  self->num_slices = 0;
+  g_array_set_size (self->entry_point_offsets, 0);
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_ensure_output_buffer (GstV4l2CodecH265Dec * self,
+    GstVideoCodecFrame * frame)
+{
+  GstBuffer *buffer;
+  GstFlowReturn flow_ret;
+
+  if (frame->output_buffer)
+    return TRUE;
+
+  flow_ret = gst_buffer_pool_acquire_buffer (GST_BUFFER_POOL (self->src_pool),
+      &buffer, NULL);
+  if (flow_ret != GST_FLOW_OK) {
+    if (flow_ret == GST_FLOW_FLUSHING)
+      GST_DEBUG_OBJECT (self, "Frame decoding aborted, we are flushing.");
+    else
+      GST_ELEMENT_ERROR (self, RESOURCE, WRITE,
+          ("No more picture buffer available."), (NULL));
+    return FALSE;
+  }
+
+  frame->output_buffer = buffer;
+  return TRUE;
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_submit_bitstream (GstV4l2CodecH265Dec * self,
+    GstH265Picture * picture, guint flags)
+{
+  GstV4l2Request *prev_request, *request = NULL;
+  gsize bytesused;
+  gboolean ret = FALSE;
+  gint num_controls = 0;
+
+  /* *INDENT-OFF* */
+  /* Reserve space for controls */
+  struct v4l2_ext_control control[] = {
+    { }, /* SPS */
+    { }, /* PPS */
+    { }, /* DECODE_PARAMS */
+    { }, /* SLICE_PARAMS */
+    { }, /* SCALING_MATRIX */
+    { }, /* ENTRY_POINT_OFFSETS */
+  };
+  /* *INDENT-ON* */
+
+  prev_request = gst_h265_picture_get_user_data (picture);
+
+  bytesused = self->bitstream_map.size;
+  gst_memory_unmap (self->bitstream, &self->bitstream_map);
+  self->bitstream_map = (GstMapInfo) GST_MAP_INFO_INIT;
+  gst_memory_resize (self->bitstream, 0, bytesused);
+
+  if (prev_request) {
+    request = gst_v4l2_decoder_alloc_sub_request (self->decoder, prev_request,
+        self->bitstream);
+  } else {
+    GstVideoCodecFrame *frame;
+
+    frame = gst_video_decoder_get_frame (GST_VIDEO_DECODER (self),
+        picture->system_frame_number);
+    g_return_val_if_fail (frame, FALSE);
+
+    if (!gst_v4l2_codec_h265_dec_ensure_output_buffer (self, frame))
+      goto done;
+
+    request = gst_v4l2_decoder_alloc_request (self->decoder,
+        picture->system_frame_number, self->bitstream, frame->output_buffer);
+
+    gst_video_codec_frame_unref (frame);
+  }
+
+  if (!request) {
+    GST_ELEMENT_ERROR (self, RESOURCE, NO_SPACE_LEFT,
+        ("Failed to allocate a media request object."), (NULL));
+    goto done;
+  }
+
+  if (self->need_sequence) {
+    control[num_controls].id = V4L2_CID_STATELESS_HEVC_SPS;
+    control[num_controls].ptr = &self->sps;
+    control[num_controls].size = sizeof (self->sps);
+    num_controls++;
+    self->need_sequence = FALSE;
+  }
+
+  if (self->first_slice) {
+    control[num_controls].id = V4L2_CID_STATELESS_HEVC_PPS;
+    control[num_controls].ptr = &self->pps;
+    control[num_controls].size = sizeof (self->pps);
+    num_controls++;
+
+    if (self->support_scaling_matrix) {
+      control[num_controls].id = V4L2_CID_STATELESS_HEVC_SCALING_MATRIX;
+      control[num_controls].ptr = &self->scaling_matrix;
+      control[num_controls].size = sizeof (self->scaling_matrix);
+      num_controls++;
+    }
+
+    control[num_controls].id = V4L2_CID_STATELESS_HEVC_DECODE_PARAMS;
+    control[num_controls].ptr = &self->decode_params;
+    control[num_controls].size = sizeof (self->decode_params);
+    num_controls++;
+
+    self->first_slice = FALSE;
+  }
+
+  /* Slice parameters are only filled for slice based or frame based with
+   * slices decoders */
+  if (self->num_slices && !is_frame_based (self)) {
+    control[num_controls].id = V4L2_CID_STATELESS_HEVC_SLICE_PARAMS;
+    control[num_controls].ptr = self->slice_params->data;
+    control[num_controls].size = g_array_get_element_size (self->slice_params)
+        * self->num_slices;
+    num_controls++;
+
+    if (self->support_entry_point_offsets && self->entry_point_offsets->len) {
+      control[num_controls].id = V4L2_CID_STATELESS_HEVC_ENTRY_POINT_OFFSETS;
+      control[num_controls].ptr = self->entry_point_offsets->data;
+      control[num_controls].size =
+          g_array_get_element_size (self->entry_point_offsets)
+          * self->entry_point_offsets->len;
+      num_controls++;
+    }
+  }
+
+  if (num_controls > G_N_ELEMENTS (control))
+    g_error ("Set too many controls, increase control[] size");
+
+  if (!gst_v4l2_decoder_set_controls (self->decoder, request, control,
+          num_controls)) {
+    GST_ELEMENT_ERROR (self, RESOURCE, WRITE,
+        ("Driver did not accept the bitstream parameters."), (NULL));
+    goto done;
+  }
+
+  if (!gst_v4l2_request_queue (request, flags)) {
+    GST_ELEMENT_ERROR (self, RESOURCE, WRITE,
+        ("Driver did not accept the decode request."), (NULL));
+    goto done;
+  }
+
+  gst_h265_picture_set_user_data (picture, g_steal_pointer (&request),
+      (GDestroyNotify) gst_v4l2_request_unref);
+  ret = TRUE;
+
+done:
+  if (request)
+    gst_v4l2_request_unref (request);
+
+  gst_v4l2_codec_h265_dec_reset_picture (self);
+
+  return ret;
+}
+
+static GstFlowReturn
+gst_v4l2_codec_h265_dec_decode_slice (GstH265Decoder * decoder,
+    GstH265Picture * picture, GstH265Slice * slice,
+    GArray * ref_pic_list0, GArray * ref_pic_list1)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  gsize sc_off = 0;
+  gsize nal_size;
+  guint8 *bitstream_data;
+
+  if (!is_frame_based (self)) {
+    if (is_slice_based (self)) {
+      if (self->bitstream_map.size) {
+        /* In slice mode, we submit the pending slice asking the accelerator to
+         * hold on the picture */
+        if (!gst_v4l2_codec_h265_dec_submit_bitstream (self, picture,
+                V4L2_BUF_FLAG_M2M_HOLD_CAPTURE_BUF)
+            || !gst_v4l2_codec_h265_dec_ensure_bitstream (self))
+          return GST_FLOW_ERROR;
+      }
+    }
+    /* in frame based mode with slices, we need to provide the required data for
+     * the whole frame, therefore we don't submit the bitstream here */
+    gst_v4l2_codec_h265_dec_fill_slice_params (self, slice, picture);
+    gst_v4l2_codec_h265_dec_fill_references (self, ref_pic_list0,
+        ref_pic_list1);
+  }
+
+  /* if it is the first slice segment provide the short and long term
+   * reference pictures set size */
+  if (slice->header.first_slice_segment_in_pic_flag) {
+    struct v4l2_ctrl_hevc_decode_params *decode_params = &self->decode_params;
+
+    decode_params->short_term_ref_pic_set_size =
+        slice->header.short_term_ref_pic_set_size;
+    decode_params->long_term_ref_pic_set_size =
+        slice->header.long_term_ref_pic_set_size;
+  }
+
+  bitstream_data = self->bitstream_map.data + self->bitstream_map.size;
+
+  if (needs_start_codes (self))
+    sc_off = 3;
+  nal_size = sc_off + slice->nalu.size;
+
+  if (self->bitstream_map.size + nal_size > self->bitstream_map.maxsize) {
+    GST_ELEMENT_ERROR (decoder, RESOURCE, NO_SPACE_LEFT,
+        ("Not enough space to send all slice of an H265 frame."), (NULL));
+    return GST_FLOW_ERROR;
+  }
+
+  if (needs_start_codes (self)) {
+    bitstream_data[0] = 0x00;
+    bitstream_data[1] = 0x00;
+    bitstream_data[2] = 0x01;
+  }
+
+  memcpy (bitstream_data + sc_off, slice->nalu.data + slice->nalu.offset,
+      slice->nalu.size);
+  self->bitstream_map.size += nal_size;
+
+  return GST_FLOW_OK;
+}
+
+static GstFlowReturn
+gst_v4l2_codec_h265_dec_end_picture (GstH265Decoder * decoder,
+    GstH265Picture * picture)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+
+  if (!gst_v4l2_codec_h265_dec_submit_bitstream (self, picture, 0))
+    return GST_FLOW_ERROR;
+
+  return GST_FLOW_OK;
+}
+
+static guint
+gst_v4l2_codec_h265_dec_get_preferred_output_delay (GstH265Decoder * decoder,
+    gboolean is_live)
+{
+
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+  guint delay;
+
+  if (is_live)
+    delay = 0;
+  else
+    delay = 1;
+
+  gst_v4l2_decoder_set_render_delay (self->decoder, delay);
+  return delay;
+}
+
+static void
+gst_v4l2_codec_h265_dec_set_flushing (GstV4l2CodecH265Dec * self,
+    gboolean flushing)
+{
+  if (self->sink_allocator)
+    gst_v4l2_codec_allocator_set_flushing (self->sink_allocator, flushing);
+  if (self->src_allocator)
+    gst_v4l2_codec_allocator_set_flushing (self->src_allocator, flushing);
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_flush (GstVideoDecoder * decoder)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+
+  GST_DEBUG_OBJECT (self, "Flushing decoder state.");
+
+  gst_v4l2_decoder_flush (self->decoder);
+  gst_v4l2_codec_h265_dec_set_flushing (self, FALSE);
+
+  return GST_VIDEO_DECODER_CLASS (parent_class)->flush (decoder);
+}
+
+static gboolean
+gst_v4l2_codec_h265_dec_sink_event (GstVideoDecoder * decoder, GstEvent * event)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder);
+
+  switch (GST_EVENT_TYPE (event)) {
+    case GST_EVENT_FLUSH_START:
+      GST_DEBUG_OBJECT (self, "flush start");
+      gst_v4l2_codec_h265_dec_set_flushing (self, TRUE);
+      break;
+    default:
+      break;
+  }
+
+  return GST_VIDEO_DECODER_CLASS (parent_class)->sink_event (decoder, event);
+}
+
+static GstStateChangeReturn
+gst_v4l2_codec_h265_dec_change_state (GstElement * element,
+    GstStateChange transition)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (element);
+
+  if (transition == GST_STATE_CHANGE_PAUSED_TO_READY)
+    gst_v4l2_codec_h265_dec_set_flushing (self, TRUE);
+
+  return GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
+}
+
+static void
+gst_v4l2_codec_h265_dec_set_property (GObject * object, guint prop_id,
+    const GValue * value, GParamSpec * pspec)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (object);
+  GObject *dec = G_OBJECT (self->decoder);
+
+  switch (prop_id) {
+    default:
+      gst_v4l2_decoder_set_property (dec, prop_id - PROP_LAST, value, pspec);
+      break;
+  }
+}
+
+static void
+gst_v4l2_codec_h265_dec_get_property (GObject * object, guint prop_id,
+    GValue * value, GParamSpec * pspec)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (object);
+  GObject *dec = G_OBJECT (self->decoder);
+
+  switch (prop_id) {
+    default:
+      gst_v4l2_decoder_get_property (dec, prop_id - PROP_LAST, value, pspec);
+      break;
+  }
+}
+
+static void
+gst_v4l2_codec_h265_dec_init (GstV4l2CodecH265Dec * self)
+{
+}
+
+static void
+gst_v4l2_codec_h265_dec_subinit (GstV4l2CodecH265Dec * self,
+    GstV4l2CodecH265DecClass * klass)
+{
+  self->decoder = gst_v4l2_decoder_new (klass->device);
+  gst_video_info_init (&self->vinfo);
+  self->slice_params = g_array_sized_new (FALSE, TRUE,
+      sizeof (struct v4l2_ctrl_hevc_slice_params), 4);
+  g_array_set_size (self->slice_params, 4);
+  self->entry_point_offsets = g_array_sized_new (FALSE, TRUE,
+      sizeof (guint32), 4);
+}
+
+static void
+gst_v4l2_codec_h265_dec_dispose (GObject * object)
+{
+  GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (object);
+
+  g_clear_object (&self->decoder);
+  g_clear_pointer (&self->slice_params, g_array_unref);
+
+  G_OBJECT_CLASS (parent_class)->dispose (object);
+}
+
+static void
+gst_v4l2_codec_h265_dec_class_init (GstV4l2CodecH265DecClass * klass)
+{
+}
+
+static void
+gst_v4l2_codec_h265_dec_subclass_init (GstV4l2CodecH265DecClass * klass,
+    GstV4l2CodecDevice * device)
+{
+  GObjectClass *gobject_class = G_OBJECT_CLASS (klass);
+  GstElementClass *element_class = GST_ELEMENT_CLASS (klass);
+  GstVideoDecoderClass *decoder_class = GST_VIDEO_DECODER_CLASS (klass);
+  GstH265DecoderClass *h265decoder_class = GST_H265_DECODER_CLASS (klass);
+
+  gobject_class->set_property = gst_v4l2_codec_h265_dec_set_property;
+  gobject_class->get_property = gst_v4l2_codec_h265_dec_get_property;
+  gobject_class->dispose = gst_v4l2_codec_h265_dec_dispose;
+
+  gst_element_class_set_static_metadata (element_class,
+      "V4L2 Stateless H.265 Video Decoder",
+      "Codec/Decoder/Video/Hardware",
+      "A V4L2 based H.265 video decoder",
+      "Nicolas Dufresne <nicolas.dufresne@collabora.com>");
+
+  gst_element_class_add_static_pad_template (element_class, &sink_template);
+  gst_element_class_add_static_pad_template (element_class, &src_template);
+  element_class->change_state =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_change_state);
+
+  decoder_class->open = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_open);
+  decoder_class->close = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_close);
+  decoder_class->stop = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_stop);
+  decoder_class->negotiate =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_negotiate);
+  decoder_class->decide_allocation =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_decide_allocation);
+  decoder_class->flush = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_flush);
+  decoder_class->sink_event =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_sink_event);
+
+  h265decoder_class->new_sequence =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_new_sequence);
+  h265decoder_class->output_picture =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_output_picture);
+  h265decoder_class->start_picture =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_start_picture);
+  h265decoder_class->decode_slice =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_decode_slice);
+  h265decoder_class->end_picture =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_end_picture);
+  h265decoder_class->get_preferred_output_delay =
+      GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_get_preferred_output_delay);
+
+  klass->device = device;
+  gst_v4l2_decoder_install_properties (gobject_class, PROP_LAST, device);
+}
+
+void
+gst_v4l2_codec_h265_dec_register (GstPlugin * plugin, GstV4l2Decoder * decoder,
+    GstV4l2CodecDevice * device, guint rank)
+{
+  GstCaps *src_caps;
+  guint version;
+
+  GST_DEBUG_CATEGORY_INIT (v4l2_h265dec_debug, "v4l2codecs-h265dec", 0,
+      "V4L2 stateless h265 decoder");
+
+  if (!gst_v4l2_decoder_set_sink_fmt (decoder, V4L2_PIX_FMT_HEVC_SLICE,
+          320, 240, 8))
+    return;
+  src_caps = gst_v4l2_decoder_enum_src_formats (decoder);
+
+  if (gst_caps_is_empty (src_caps)) {
+    GST_WARNING ("Not registering H265 decoder since it produces no "
+        "supported format");
+    goto done;
+  }
+
+  version = gst_v4l2_decoder_get_version (decoder);
+  if (version < V4L2_MIN_KERNEL_VERSION)
+    GST_WARNING ("V4L2 API v%u.%u too old, at least v%u.%u required",
+        (version >> 16) & 0xff, (version >> 8) & 0xff,
+        V4L2_MIN_KERNEL_VER_MAJOR, V4L2_MIN_KERNEL_VER_MINOR);
+
+  gst_v4l2_decoder_register (plugin, GST_TYPE_V4L2_CODEC_H265_DEC,
+      (GClassInitFunc) gst_v4l2_codec_h265_dec_subclass_init,
+      gst_mini_object_ref (GST_MINI_OBJECT (device)),
+      (GInstanceInitFunc) gst_v4l2_codec_h265_dec_subinit,
+      "v4l2sl%sh265dec", device, rank, NULL);
+
+done:
+  gst_caps_unref (src_caps);
+}
diff --git a/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.h b/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.h
new file mode 100644 (file)
index 0000000..2a73435
--- /dev/null
@@ -0,0 +1,55 @@
+/* GStreamer
+ * Copyright (C) 2020 Nicolas Dufresne <nicolas.dufresne@collabora.com>
+ * Copyright (C) 2020 Safran Passenger Innovations LLC
+ *
+ * This library is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Library General Public
+ * License as published by the Free Software Foundation; either
+ * version 2 of the License, or (at your option) any later version.
+ *
+ * This library is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
+ * Library General Public License for more details.
+ *
+ * You should have received a copy of the GNU Library General Public
+ * License along with this library; if not, write to the
+ * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
+ * Boston, MA 02110-1301, USA.
+ */
+
+#ifndef __GST_V4L2_CODEC_H265_DEC_H__
+#define __GST_V4L2_CODEC_H265_DEC_H__
+
+#define GST_USE_UNSTABLE_API
+#include <gst/codecs/gsth265decoder.h>
+
+#include "gstv4l2decoder.h"
+
+G_BEGIN_DECLS
+
+#define GST_TYPE_V4L2_CODEC_H265_DEC           (gst_v4l2_codec_h265_dec_get_type())
+#define GST_V4L2_CODEC_H265_DEC(obj)           (G_TYPE_CHECK_INSTANCE_CAST((obj),GST_TYPE_V4L2_CODEC_H265_DEC,GstV4l2CodecH265Dec))
+#define GST_V4L2_CODEC_H265_DEC_CLASS(klass)   (G_TYPE_CHECK_CLASS_CAST((klass),GST_TYPE_V4L2_CODEC_H265_DEC,GstV4l2CodecH265DecClass))
+#define GST_V4L2_CODEC_H265_DEC_GET_CLASS(obj) (G_TYPE_INSTANCE_GET_CLASS ((obj), GST_TYPE_V4L2_CODEC_H265_DEC, GstV4l2CodecH265DecClass))
+#define GST_IS_V4L2_CODEC_H265_DEC(obj)        (G_TYPE_CHECK_INSTANCE_TYPE((obj),GST_TYPE_V4L2_CODEC_H265_DEC))
+#define GST_IS_V4L2_CODEC_H265_DEC_CLASS(obj)  (G_TYPE_CHECK_CLASS_TYPE((klass),GST_TYPE_V4L2_CODEC_H265_DEC))
+
+typedef struct _GstV4l2CodecH265Dec GstV4l2CodecH265Dec;
+typedef struct _GstV4l2CodecH265DecClass GstV4l2CodecH265DecClass;
+
+struct _GstV4l2CodecH265DecClass
+{
+  GstH265DecoderClass parent_class;
+  GstV4l2CodecDevice *device;
+};
+
+GType gst_v4l2_codec_h265_dec_get_type (void);
+void  gst_v4l2_codec_h265_dec_register (GstPlugin * plugin,
+                                        GstV4l2Decoder * decoder,
+                                        GstV4l2CodecDevice * device,
+                                        guint rank);
+
+G_END_DECLS
+
+#endif /* __GST_D3D11_H265_DEC_H__ */
index ea8c79c..6b44dca 100644 (file)
@@ -3,6 +3,7 @@ v4l2codecs_sources = [
   'gstv4l2codecallocator.c',
   'gstv4l2codecdevice.c',
   'gstv4l2codech264dec.c',
+  'gstv4l2codech265dec.c',
   'gstv4l2codecmpeg2dec.c',
   'gstv4l2codecpool.c',
   'gstv4l2codecvp8dec.c',
index 7d0796e..ac83e1b 100644 (file)
@@ -24,6 +24,7 @@
 
 #include "gstv4l2codecdevice.h"
 #include "gstv4l2codech264dec.h"
+#include "gstv4l2codech265dec.h"
 #include "gstv4l2codecmpeg2dec.h"
 #include "gstv4l2codecvp8dec.h"
 #include "gstv4l2codecvp9dec.h"
@@ -54,6 +55,13 @@ register_video_decoder (GstPlugin * plugin, GstV4l2CodecDevice * device)
         gst_v4l2_codec_h264_dec_register (plugin, decoder, device,
             GST_RANK_PRIMARY + 1);
         break;
+      case V4L2_PIX_FMT_HEVC_SLICE:
+        GST_INFO_OBJECT (decoder, "Registering %s as H265 Decoder",
+            device->name);
+        gst_v4l2_codec_h265_dec_register (plugin, decoder, device,
+            GST_RANK_PRIMARY + 1);
+        break;
+
       case V4L2_PIX_FMT_VP8_FRAME:
         GST_INFO_OBJECT (decoder, "Registering %s as VP8 Decoder",
             device->name);