2 * Copyright (C) 2014-2017 SUMOMO Computer Association
3 * Authors Ayaka <ayaka@soulik.info>
4 * Copyright (C) 2017 Collabora Ltd.
5 * Author: Nicolas Dufresne <nicolas.dufresne@collabora.com>
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Library General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Library General Public License for more details.
17 * You should have received a copy of the GNU Library General Public
18 * License along with this library; if not, write to the
19 * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor,
20 * Boston, MA 02110-1301, USA.
34 #include "gstv4l2object.h"
35 #include "gstv4l2videoenc.h"
38 #include <gst/gst-i18n-plugin.h>
40 GST_DEBUG_CATEGORY_STATIC (gst_v4l2_video_enc_debug);
41 #define GST_CAT_DEFAULT gst_v4l2_video_enc_debug
48 } GstV4l2VideoEncCData;
53 V4L2_STD_OBJECT_PROPS,
56 #define gst_v4l2_video_enc_parent_class parent_class
57 G_DEFINE_ABSTRACT_TYPE (GstV4l2VideoEnc, gst_v4l2_video_enc,
58 GST_TYPE_VIDEO_ENCODER);
61 gst_v4l2_video_enc_set_property (GObject * object,
62 guint prop_id, const GValue * value, GParamSpec * pspec)
64 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (object);
67 case PROP_OUTPUT_IO_MODE:
68 gst_v4l2_object_set_property_helper (self->v4l2output,
69 prop_id, value, pspec);
71 case PROP_CAPTURE_IO_MODE:
72 gst_v4l2_object_set_property_helper (self->v4l2capture,
73 prop_id, value, pspec);
76 /* By default, only set on output */
78 if (!gst_v4l2_object_set_property_helper (self->v4l2output,
79 prop_id, value, pspec)) {
80 G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
87 gst_v4l2_video_enc_get_property (GObject * object,
88 guint prop_id, GValue * value, GParamSpec * pspec)
90 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (object);
93 case PROP_OUTPUT_IO_MODE:
94 gst_v4l2_object_get_property_helper (self->v4l2output,
95 prop_id, value, pspec);
97 case PROP_CAPTURE_IO_MODE:
98 gst_v4l2_object_get_property_helper (self->v4l2output,
99 PROP_IO_MODE, value, pspec);
102 /* By default read from output */
104 if (!gst_v4l2_object_get_property_helper (self->v4l2output,
105 prop_id, value, pspec)) {
106 G_OBJECT_WARN_INVALID_PROPERTY_ID (object, prop_id, pspec);
113 gst_v4l2_video_enc_open (GstVideoEncoder * encoder)
115 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
118 GST_DEBUG_OBJECT (self, "Opening");
120 if (!gst_v4l2_object_open (self->v4l2output))
123 if (!gst_v4l2_object_open_shared (self->v4l2capture, self->v4l2output))
126 self->probed_sinkcaps = gst_v4l2_object_probe_caps (self->v4l2output,
127 gst_v4l2_object_get_raw_caps ());
129 if (gst_caps_is_empty (self->probed_sinkcaps))
132 codec_caps = gst_pad_get_pad_template_caps (encoder->srcpad);
133 self->probed_srccaps = gst_v4l2_object_probe_caps (self->v4l2capture,
135 gst_caps_unref (codec_caps);
137 if (gst_caps_is_empty (self->probed_srccaps))
138 goto no_encoded_format;
143 GST_ELEMENT_ERROR (self, RESOURCE, SETTINGS,
144 (_("Encoder on device %s has no supported output format"),
145 self->v4l2output->videodev), (NULL));
150 GST_ELEMENT_ERROR (self, RESOURCE, SETTINGS,
151 (_("Encoder on device %s has no supported input format"),
152 self->v4l2output->videodev), (NULL));
156 if (GST_V4L2_IS_OPEN (self->v4l2output))
157 gst_v4l2_object_close (self->v4l2output);
159 if (GST_V4L2_IS_OPEN (self->v4l2capture))
160 gst_v4l2_object_close (self->v4l2capture);
162 gst_caps_replace (&self->probed_srccaps, NULL);
163 gst_caps_replace (&self->probed_sinkcaps, NULL);
169 gst_v4l2_video_enc_close (GstVideoEncoder * encoder)
171 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
173 GST_DEBUG_OBJECT (self, "Closing");
175 gst_v4l2_object_close (self->v4l2output);
176 gst_v4l2_object_close (self->v4l2capture);
177 gst_caps_replace (&self->probed_srccaps, NULL);
178 gst_caps_replace (&self->probed_sinkcaps, NULL);
184 gst_v4l2_video_enc_start (GstVideoEncoder * encoder)
186 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
188 GST_DEBUG_OBJECT (self, "Starting");
190 gst_v4l2_object_unlock (self->v4l2output);
191 g_atomic_int_set (&self->active, TRUE);
192 self->output_flow = GST_FLOW_OK;
198 gst_v4l2_video_enc_stop (GstVideoEncoder * encoder)
200 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
202 GST_DEBUG_OBJECT (self, "Stopping");
204 gst_v4l2_object_unlock (self->v4l2output);
205 gst_v4l2_object_unlock (self->v4l2capture);
207 /* Wait for capture thread to stop */
208 gst_pad_stop_task (encoder->srcpad);
210 GST_VIDEO_ENCODER_STREAM_LOCK (encoder);
211 self->output_flow = GST_FLOW_OK;
212 GST_VIDEO_ENCODER_STREAM_UNLOCK (encoder);
214 /* Should have been flushed already */
215 g_assert (g_atomic_int_get (&self->active) == FALSE);
216 g_assert (g_atomic_int_get (&self->processing) == FALSE);
218 gst_v4l2_object_stop (self->v4l2output);
219 gst_v4l2_object_stop (self->v4l2capture);
221 if (self->v4l2output->pool)
222 gst_v4l2_buffer_pool_flush (self->v4l2output->pool);
223 if (self->v4l2capture->pool)
224 gst_v4l2_buffer_pool_flush (self->v4l2capture->pool);
226 if (self->input_state) {
227 gst_video_codec_state_unref (self->input_state);
228 self->input_state = NULL;
231 GST_DEBUG_OBJECT (self, "Stopped");
237 gst_v4l2_encoder_cmd (GstV4l2Object * v4l2object, guint cmd, guint flags)
239 struct v4l2_encoder_cmd ecmd = { 0, };
241 GST_DEBUG_OBJECT (v4l2object->element,
242 "sending v4l2 encoder command %u with flags %u", cmd, flags);
244 if (!GST_V4L2_IS_OPEN (v4l2object))
249 if (v4l2object->ioctl (v4l2object->video_fd, VIDIOC_ENCODER_CMD, &ecmd) < 0)
255 if (errno == ENOTTY) {
256 GST_INFO_OBJECT (v4l2object->element,
257 "Failed to send encoder command %u with flags %u for '%s'. (%s)",
258 cmd, flags, v4l2object->videodev, g_strerror (errno));
260 GST_ERROR_OBJECT (v4l2object->element,
261 "Failed to send encoder command %u with flags %u for '%s'. (%s)",
262 cmd, flags, v4l2object->videodev, g_strerror (errno));
268 gst_v4l2_video_enc_finish (GstVideoEncoder * encoder)
270 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
271 GstFlowReturn ret = GST_FLOW_OK;
273 if (gst_pad_get_task_state (encoder->srcpad) != GST_TASK_STARTED)
276 GST_DEBUG_OBJECT (self, "Finishing encoding");
278 /* drop the stream lock while draining, so remaining buffers can be
279 * pushed from the src pad task thread */
280 GST_VIDEO_ENCODER_STREAM_UNLOCK (encoder);
282 if (gst_v4l2_encoder_cmd (self->v4l2capture, V4L2_ENC_CMD_STOP, 0)) {
283 GstTask *task = encoder->srcpad->task;
285 /* Wait for the task to be drained */
286 GST_OBJECT_LOCK (task);
287 while (GST_TASK_STATE (task) == GST_TASK_STARTED)
288 GST_TASK_WAIT (task);
289 GST_OBJECT_UNLOCK (task);
290 ret = GST_FLOW_FLUSHING;
293 /* and ensure the processing thread has stopped in case another error
295 gst_v4l2_object_unlock (self->v4l2capture);
296 gst_pad_stop_task (encoder->srcpad);
297 GST_VIDEO_ENCODER_STREAM_LOCK (encoder);
299 if (ret == GST_FLOW_FLUSHING)
300 ret = self->output_flow;
302 GST_DEBUG_OBJECT (encoder, "Done draining buffers");
309 gst_v4l2_video_enc_set_format (GstVideoEncoder * encoder,
310 GstVideoCodecState * state)
313 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
314 GstV4l2Error error = GST_V4L2_ERROR_INIT;
316 GstVideoCodecState *output;
318 GST_DEBUG_OBJECT (self, "Setting format: %" GST_PTR_FORMAT, state->caps);
320 if (self->input_state) {
321 if (gst_v4l2_object_caps_equal (self->v4l2output, state->caps)) {
322 GST_DEBUG_OBJECT (self, "Compatible caps");
326 if (gst_v4l2_video_enc_finish (encoder) != GST_FLOW_OK)
329 gst_video_codec_state_unref (self->input_state);
330 self->input_state = NULL;
333 outcaps = gst_pad_get_pad_template_caps (encoder->srcpad);
334 outcaps = gst_caps_make_writable (outcaps);
335 output = gst_video_encoder_set_output_state (encoder, outcaps, state);
336 gst_video_codec_state_unref (output);
338 if (!gst_video_encoder_negotiate (encoder))
341 if (!gst_v4l2_object_set_format (self->v4l2output, state->caps, &error)) {
342 gst_v4l2_error (self, &error);
346 self->input_state = gst_video_codec_state_ref (state);
348 GST_DEBUG_OBJECT (self, "output caps: %" GST_PTR_FORMAT, state->caps);
354 gst_v4l2_video_enc_flush (GstVideoEncoder * encoder)
356 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
358 GST_DEBUG_OBJECT (self, "Flushing");
360 /* Ensure the processing thread has stopped for the reverse playback
362 if (g_atomic_int_get (&self->processing)) {
363 GST_VIDEO_ENCODER_STREAM_UNLOCK (encoder);
365 gst_v4l2_object_unlock_stop (self->v4l2output);
366 gst_v4l2_object_unlock_stop (self->v4l2capture);
367 gst_pad_stop_task (encoder->srcpad);
369 GST_VIDEO_ENCODER_STREAM_UNLOCK (encoder);
373 self->output_flow = GST_FLOW_OK;
375 gst_v4l2_object_unlock_stop (self->v4l2output);
376 gst_v4l2_object_unlock_stop (self->v4l2capture);
381 struct ProfileLevelCtx
383 GstV4l2VideoEnc *self;
384 const gchar *profile;
389 get_string_list (GstStructure * s, const gchar * field, GQueue * queue)
393 value = gst_structure_get_value (s, field);
398 if (GST_VALUE_HOLDS_LIST (value)) {
401 if (gst_value_list_get_size (value) == 0)
404 for (i = 0; i < gst_value_list_get_size (value); i++) {
405 const GValue *item = gst_value_list_get_value (value, i);
407 if (G_VALUE_HOLDS_STRING (item))
408 g_queue_push_tail (queue, g_value_dup_string (item));
410 } else if (G_VALUE_HOLDS_STRING (value)) {
411 g_queue_push_tail (queue, g_value_dup_string (value));
418 negotiate_profile_and_level (GstCapsFeatures * features, GstStructure * s,
421 struct ProfileLevelCtx *ctx = user_data;
422 GstV4l2VideoEncClass *klass = GST_V4L2_VIDEO_ENC_GET_CLASS (ctx->self);
423 GstV4l2Object *v4l2object = GST_V4L2_VIDEO_ENC (ctx->self)->v4l2output;
424 GQueue profiles = G_QUEUE_INIT;
425 GQueue levels = G_QUEUE_INIT;
426 gboolean failed = FALSE;
428 if (klass->profile_cid && get_string_list (s, "profile", &profiles)) {
431 for (l = profiles.head; l; l = l->next) {
432 struct v4l2_control control = { 0, };
434 const gchar *profile = l->data;
436 GST_TRACE_OBJECT (ctx->self, "Trying profile %s", profile);
438 control.id = klass->profile_cid;
439 control.value = v4l2_profile = klass->profile_from_string (profile);
441 if (control.value < 0)
444 if (v4l2object->ioctl (v4l2object->video_fd, VIDIOC_S_CTRL, &control) < 0) {
445 GST_WARNING_OBJECT (ctx->self, "Failed to set %s profile: '%s'",
446 klass->codec_name, g_strerror (errno));
450 profile = klass->profile_to_string (control.value);
452 if (control.value == v4l2_profile) {
453 ctx->profile = profile;
457 if (g_list_find_custom (l, profile, g_str_equal)) {
458 ctx->profile = profile;
463 if (profiles.length && !ctx->profile)
466 g_queue_foreach (&profiles, (GFunc) g_free, NULL);
467 g_queue_clear (&profiles);
470 if (!failed && klass->level_cid && get_string_list (s, "level", &levels)) {
473 for (l = levels.head; l; l = l->next) {
474 struct v4l2_control control = { 0, };
476 const gchar *level = l->data;
478 GST_TRACE_OBJECT (ctx->self, "Trying level %s", level);
480 control.id = klass->level_cid;
481 control.value = v4l2_level = klass->level_from_string (level);
483 if (control.value < 0)
486 if (v4l2object->ioctl (v4l2object->video_fd, VIDIOC_S_CTRL, &control) < 0) {
487 GST_WARNING_OBJECT (ctx->self, "Failed to set %s level: '%s'",
488 klass->codec_name, g_strerror (errno));
492 level = klass->level_to_string (control.value);
494 if (control.value == v4l2_level) {
499 if (g_list_find_custom (l, level, g_str_equal)) {
505 if (levels.length && !ctx->level)
508 g_queue_foreach (&levels, (GFunc) g_free, NULL);
509 g_queue_clear (&levels);
512 /* If it failed, we continue */
517 gst_v4l2_video_enc_negotiate (GstVideoEncoder * encoder)
519 GstV4l2VideoEncClass *klass = GST_V4L2_VIDEO_ENC_GET_CLASS (encoder);
520 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
521 GstV4l2Object *v4l2object = self->v4l2output;
522 GstCaps *allowed_caps;
523 struct ProfileLevelCtx ctx = { self, NULL, NULL };
524 GstVideoCodecState *state;
527 GST_DEBUG_OBJECT (self, "Negotiating %s profile and level.",
530 allowed_caps = gst_pad_get_allowed_caps (GST_VIDEO_ENCODER_SRC_PAD (encoder));
534 if (gst_caps_is_empty (allowed_caps))
537 allowed_caps = gst_caps_make_writable (allowed_caps);
539 /* negotiate_profile_and_level() will return TRUE on failure to keep
540 * iterating, if gst_caps_foreach() returns TRUE it means there was no
541 * compatible profile and level in any of the structure */
542 if (gst_caps_foreach (allowed_caps, negotiate_profile_and_level, &ctx)) {
543 goto no_profile_level;
547 if (klass->profile_cid && !ctx.profile) {
548 struct v4l2_control control = { 0, };
550 control.id = klass->profile_cid;
552 if (v4l2object->ioctl (v4l2object->video_fd, VIDIOC_G_CTRL, &control) < 0)
555 ctx.profile = klass->profile_to_string (control.value);
558 if (klass->level_cid && !ctx.level) {
559 struct v4l2_control control = { 0, };
561 control.id = klass->level_cid;
563 if (v4l2object->ioctl (v4l2object->video_fd, VIDIOC_G_CTRL, &control) < 0)
566 ctx.level = klass->level_to_string (control.value);
569 GST_DEBUG_OBJECT (self, "Selected %s profile %s at level %s",
570 klass->codec_name, ctx.profile, ctx.level);
572 state = gst_video_encoder_get_output_state (encoder);
573 s = gst_caps_get_structure (state->caps, 0);
575 if (klass->profile_cid)
576 gst_structure_set (s, "profile", G_TYPE_STRING, ctx.profile, NULL);
578 if (klass->level_cid)
579 gst_structure_set (s, "level", G_TYPE_STRING, ctx.level, NULL);
581 if (!GST_VIDEO_ENCODER_CLASS (parent_class)->negotiate (encoder))
584 if (!gst_buffer_pool_set_active (GST_BUFFER_POOL (self->v4l2capture->pool),
586 GST_WARNING_OBJECT (self, "Could not activate capture buffer pool.");
593 GST_WARNING_OBJECT (self, "Failed to get %s profile and level: '%s'",
594 klass->codec_name, g_strerror (errno));
598 GST_WARNING_OBJECT (self, "No compatible level and profile in caps: %"
599 GST_PTR_FORMAT, allowed_caps);
604 gst_caps_unref (allowed_caps);
608 static GstVideoCodecFrame *
609 gst_v4l2_video_enc_get_oldest_frame (GstVideoEncoder * encoder)
611 GstVideoCodecFrame *frame = NULL;
615 frames = gst_video_encoder_get_frames (encoder);
617 for (l = frames; l != NULL; l = l->next) {
618 GstVideoCodecFrame *f = l->data;
620 if (!frame || frame->pts > f->pts)
627 GST_LOG_OBJECT (encoder,
628 "Oldest frame is %d %" GST_TIME_FORMAT
629 " and %d frames left",
630 frame->system_frame_number, GST_TIME_ARGS (frame->pts), count - 1);
631 gst_video_codec_frame_ref (frame);
634 g_list_free_full (frames, (GDestroyNotify) gst_video_codec_frame_unref);
640 gst_v4l2_video_enc_loop (GstVideoEncoder * encoder)
642 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
643 GstVideoCodecFrame *frame;
644 GstBuffer *buffer = NULL;
647 GST_LOG_OBJECT (encoder, "Allocate output buffer");
649 buffer = gst_video_encoder_allocate_output_buffer (encoder,
650 self->v4l2capture->info.size);
652 if (NULL == buffer) {
653 ret = GST_FLOW_FLUSHING;
658 /* FIXME Check if buffer isn't the last one here */
660 GST_LOG_OBJECT (encoder, "Process output buffer");
662 gst_v4l2_buffer_pool_process (GST_V4L2_BUFFER_POOL
663 (self->v4l2capture->pool), &buffer);
665 if (ret != GST_FLOW_OK)
668 frame = gst_v4l2_video_enc_get_oldest_frame (encoder);
671 frame->output_buffer = buffer;
673 ret = gst_video_encoder_finish_frame (encoder, frame);
675 if (ret != GST_FLOW_OK)
678 GST_WARNING_OBJECT (encoder, "Encoder is producing too many buffers");
679 gst_buffer_unref (buffer);
685 GST_DEBUG_OBJECT (encoder, "Leaving output thread");
687 gst_buffer_replace (&buffer, NULL);
688 self->output_flow = ret;
689 g_atomic_int_set (&self->processing, FALSE);
690 gst_v4l2_object_unlock (self->v4l2output);
691 gst_pad_pause_task (encoder->srcpad);
695 gst_v4l2_video_enc_loop_stopped (GstV4l2VideoEnc * self)
697 if (g_atomic_int_get (&self->processing)) {
698 GST_DEBUG_OBJECT (self, "Early stop of encoding thread");
699 self->output_flow = GST_FLOW_FLUSHING;
700 g_atomic_int_set (&self->processing, FALSE);
703 GST_DEBUG_OBJECT (self, "Encoding task destroyed: %s",
704 gst_flow_get_name (self->output_flow));
709 gst_v4l2_video_enc_handle_frame (GstVideoEncoder * encoder,
710 GstVideoCodecFrame * frame)
712 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
713 GstFlowReturn ret = GST_FLOW_OK;
715 GST_DEBUG_OBJECT (self, "Handling frame %d", frame->system_frame_number);
717 if (G_UNLIKELY (!g_atomic_int_get (&self->active)))
720 if (gst_pad_get_task_state (GST_VIDEO_DECODER_SRC_PAD (self)) ==
722 GstBufferPool *pool = GST_BUFFER_POOL (self->v4l2output->pool);
724 /* It possible that the processing thread stopped due to an error */
725 if (self->output_flow != GST_FLOW_OK &&
726 self->output_flow != GST_FLOW_FLUSHING) {
727 GST_DEBUG_OBJECT (self, "Processing loop stopped with error, leaving");
728 ret = self->output_flow;
732 /* Ensure input internal pool is active */
733 if (!gst_buffer_pool_is_active (pool)) {
734 GstStructure *config = gst_buffer_pool_get_config (pool);
735 guint min = MAX (self->v4l2output->min_buffers, GST_V4L2_MIN_BUFFERS);
737 gst_buffer_pool_config_set_params (config, self->input_state->caps,
738 self->v4l2output->info.size, min, min);
740 /* There is no reason to refuse this config */
741 if (!gst_buffer_pool_set_config (pool, config))
742 goto activate_failed;
744 if (!gst_buffer_pool_set_active (pool, TRUE))
745 goto activate_failed;
748 GST_DEBUG_OBJECT (self, "Starting encoding thread");
750 /* Start the processing task, when it quits, the task will disable input
751 * processing to unlock input if draining, or prevent potential block */
752 if (!gst_pad_start_task (encoder->srcpad,
753 (GstTaskFunction) gst_v4l2_video_enc_loop, self,
754 (GDestroyNotify) gst_v4l2_video_enc_loop_stopped))
755 goto start_task_failed;
758 if (frame->input_buffer) {
759 GST_VIDEO_ENCODER_STREAM_UNLOCK (encoder);
761 gst_v4l2_buffer_pool_process (GST_V4L2_BUFFER_POOL
762 (self->v4l2output->pool), &frame->input_buffer);
763 GST_VIDEO_ENCODER_STREAM_LOCK (encoder);
765 if (ret == GST_FLOW_FLUSHING) {
766 if (gst_pad_get_task_state (GST_VIDEO_DECODER_SRC_PAD (self)) !=
768 ret = self->output_flow;
770 } else if (ret != GST_FLOW_OK) {
775 gst_video_codec_frame_unref (frame);
781 GST_ELEMENT_ERROR (self, RESOURCE, SETTINGS,
782 (_("Failed to allocate required memory.")),
783 ("Buffer pool activation failed"));
784 return GST_FLOW_ERROR;
789 ret = GST_FLOW_FLUSHING;
794 GST_ELEMENT_ERROR (self, RESOURCE, FAILED,
795 (_("Failed to start encoding thread.")), (NULL));
796 g_atomic_int_set (&self->processing, FALSE);
797 ret = GST_FLOW_ERROR;
802 GST_ELEMENT_ERROR (self, RESOURCE, FAILED,
803 (_("Failed to process frame.")),
804 ("Maybe be due to not enough memory or failing driver"));
805 ret = GST_FLOW_ERROR;
810 gst_video_encoder_finish_frame (encoder, frame);
816 gst_v4l2_video_enc_decide_allocation (GstVideoEncoder *
817 encoder, GstQuery * query)
819 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
820 GstVideoCodecState *state = gst_video_encoder_get_output_state (encoder);
821 GstV4l2Error error = GST_V4L2_ERROR_INIT;
822 GstClockTime latency;
823 gboolean ret = FALSE;
825 /* We need to set the format here, since this is called right after
826 * GstVideoEncoder have set the width, height and framerate into the state
827 * caps. These are needed by the driver to calculate the buffer size and to
828 * implement bitrate adaptation. */
829 if (!gst_v4l2_object_set_format (self->v4l2capture, state->caps, &error)) {
830 gst_v4l2_error (self, &error);
835 if (gst_v4l2_object_decide_allocation (self->v4l2capture, query)) {
836 GstVideoEncoderClass *enc_class = GST_VIDEO_ENCODER_CLASS (parent_class);
837 ret = enc_class->decide_allocation (encoder, query);
840 /* FIXME This may not be entirely correct, as encoder may keep some
841 * observation withouth delaying the encoding. Linux Media API need some
842 * more work to explicitly expressed the decoder / encoder latency. This
843 * value will then become max latency, and the reported driver latency would
844 * become the min latency. */
845 latency = self->v4l2capture->min_buffers * self->v4l2capture->duration;
846 gst_video_encoder_set_latency (encoder, latency, latency);
849 gst_video_codec_state_unref (state);
854 gst_v4l2_video_enc_propose_allocation (GstVideoEncoder *
855 encoder, GstQuery * query)
857 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
858 gboolean ret = FALSE;
860 GST_DEBUG_OBJECT (self, "called");
865 ret = gst_v4l2_object_propose_allocation (self->v4l2output, query);
868 ret = GST_VIDEO_ENCODER_CLASS (parent_class)->propose_allocation (encoder,
875 gst_v4l2_video_enc_src_query (GstVideoEncoder * encoder, GstQuery * query)
878 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
879 switch (GST_QUERY_TYPE (query)) {
880 case GST_QUERY_CAPS:{
881 GstCaps *filter, *result = NULL;
882 GstPad *pad = GST_VIDEO_ENCODER_SRC_PAD (encoder);
884 gst_query_parse_caps (query, &filter);
886 /* FIXME Try and not probe the entire encoder, but only the implement
888 if (self->probed_srccaps) {
889 GstCaps *tmpl = gst_pad_get_pad_template_caps (pad);
890 result = gst_caps_intersect (tmpl, self->probed_srccaps);
891 gst_caps_unref (tmpl);
893 result = gst_pad_get_pad_template_caps (pad);
896 GstCaps *tmp = result;
898 gst_caps_intersect_full (filter, tmp, GST_CAPS_INTERSECT_FIRST);
899 gst_caps_unref (tmp);
902 GST_DEBUG_OBJECT (self, "Returning src caps %" GST_PTR_FORMAT, result);
904 gst_query_set_caps_result (query, result);
905 gst_caps_unref (result);
910 ret = GST_VIDEO_ENCODER_CLASS (parent_class)->src_query (encoder, query);
918 gst_v4l2_video_enc_sink_query (GstVideoEncoder * encoder, GstQuery * query)
921 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
923 switch (GST_QUERY_TYPE (query)) {
924 case GST_QUERY_CAPS:{
925 GstCaps *filter, *result = NULL;
926 GstPad *pad = GST_VIDEO_ENCODER_SINK_PAD (encoder);
928 gst_query_parse_caps (query, &filter);
930 if (self->probed_sinkcaps)
931 result = gst_caps_ref (self->probed_sinkcaps);
933 result = gst_pad_get_pad_template_caps (pad);
936 GstCaps *tmp = result;
938 gst_caps_intersect_full (filter, tmp, GST_CAPS_INTERSECT_FIRST);
939 gst_caps_unref (tmp);
942 GST_DEBUG_OBJECT (self, "Returning sink caps %" GST_PTR_FORMAT, result);
944 gst_query_set_caps_result (query, result);
945 gst_caps_unref (result);
950 ret = GST_VIDEO_ENCODER_CLASS (parent_class)->sink_query (encoder, query);
958 gst_v4l2_video_enc_sink_event (GstVideoEncoder * encoder, GstEvent * event)
960 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (encoder);
963 switch (GST_EVENT_TYPE (event)) {
964 case GST_EVENT_FLUSH_START:
965 GST_DEBUG_OBJECT (self, "flush start");
966 gst_v4l2_object_unlock (self->v4l2output);
967 gst_v4l2_object_unlock (self->v4l2capture);
973 ret = GST_VIDEO_ENCODER_CLASS (parent_class)->sink_event (encoder, event);
975 switch (GST_EVENT_TYPE (event)) {
976 case GST_EVENT_FLUSH_START:
977 gst_pad_stop_task (encoder->srcpad);
978 GST_DEBUG_OBJECT (self, "flush start done");
986 static GstStateChangeReturn
987 gst_v4l2_video_enc_change_state (GstElement * element,
988 GstStateChange transition)
990 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (element);
992 if (transition == GST_STATE_CHANGE_PAUSED_TO_READY) {
993 g_atomic_int_set (&self->active, FALSE);
994 gst_v4l2_object_unlock (self->v4l2output);
995 gst_v4l2_object_unlock (self->v4l2capture);
998 return GST_ELEMENT_CLASS (parent_class)->change_state (element, transition);
1003 gst_v4l2_video_enc_dispose (GObject * object)
1005 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (object);
1007 gst_caps_replace (&self->probed_sinkcaps, NULL);
1008 gst_caps_replace (&self->probed_srccaps, NULL);
1010 G_OBJECT_CLASS (parent_class)->dispose (object);
1014 gst_v4l2_video_enc_finalize (GObject * object)
1016 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (object);
1018 gst_v4l2_object_destroy (self->v4l2capture);
1019 gst_v4l2_object_destroy (self->v4l2output);
1021 G_OBJECT_CLASS (parent_class)->finalize (object);
1026 gst_v4l2_video_enc_init (GstV4l2VideoEnc * self)
1028 /* V4L2 object are created in subinstance_init */
1032 gst_v4l2_video_enc_subinstance_init (GTypeInstance * instance, gpointer g_class)
1034 GstV4l2VideoEncClass *klass = GST_V4L2_VIDEO_ENC_CLASS (g_class);
1035 GstV4l2VideoEnc *self = GST_V4L2_VIDEO_ENC (instance);
1037 self->v4l2output = gst_v4l2_object_new (GST_ELEMENT (self),
1038 V4L2_BUF_TYPE_VIDEO_OUTPUT, klass->default_device,
1039 gst_v4l2_get_output, gst_v4l2_set_output, NULL);
1040 self->v4l2output->no_initial_format = TRUE;
1041 self->v4l2output->keep_aspect = FALSE;
1043 self->v4l2capture = gst_v4l2_object_new (GST_ELEMENT (self),
1044 V4L2_BUF_TYPE_VIDEO_CAPTURE, klass->default_device,
1045 gst_v4l2_get_input, gst_v4l2_set_input, NULL);
1046 self->v4l2capture->no_initial_format = TRUE;
1047 self->v4l2output->keep_aspect = FALSE;
1051 gst_v4l2_video_enc_class_init (GstV4l2VideoEncClass * klass)
1053 GstElementClass *element_class;
1054 GObjectClass *gobject_class;
1055 GstVideoEncoderClass *video_encoder_class;
1057 parent_class = g_type_class_peek_parent (klass);
1059 element_class = (GstElementClass *) klass;
1060 gobject_class = (GObjectClass *) klass;
1061 video_encoder_class = (GstVideoEncoderClass *) klass;
1063 GST_DEBUG_CATEGORY_INIT (gst_v4l2_video_enc_debug, "v4l2videoenc", 0,
1064 "V4L2 Video Encoder");
1066 gobject_class->dispose = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_dispose);
1067 gobject_class->finalize = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_finalize);
1068 gobject_class->set_property =
1069 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_set_property);
1070 gobject_class->get_property =
1071 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_get_property);
1073 video_encoder_class->open = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_open);
1074 video_encoder_class->close = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_close);
1075 video_encoder_class->start = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_start);
1076 video_encoder_class->stop = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_stop);
1077 video_encoder_class->finish = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_finish);
1078 video_encoder_class->flush = GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_flush);
1079 video_encoder_class->set_format =
1080 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_set_format);
1081 video_encoder_class->negotiate =
1082 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_negotiate);
1083 video_encoder_class->decide_allocation =
1084 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_decide_allocation);
1085 video_encoder_class->propose_allocation =
1086 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_propose_allocation);
1087 video_encoder_class->sink_query =
1088 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_sink_query);
1089 video_encoder_class->src_query =
1090 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_src_query);
1091 video_encoder_class->sink_event =
1092 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_sink_event);
1093 video_encoder_class->handle_frame =
1094 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_handle_frame);
1096 element_class->change_state =
1097 GST_DEBUG_FUNCPTR (gst_v4l2_video_enc_change_state);
1099 gst_v4l2_object_install_m2m_properties_helper (gobject_class);
1103 gst_v4l2_video_enc_subclass_init (gpointer g_class, gpointer data)
1105 GstV4l2VideoEncClass *klass = GST_V4L2_VIDEO_ENC_CLASS (g_class);
1106 GstElementClass *element_class = GST_ELEMENT_CLASS (g_class);
1107 GstV4l2VideoEncCData *cdata = data;
1109 klass->default_device = cdata->device;
1111 /* Note: gst_pad_template_new() take the floating ref from the caps */
1112 gst_element_class_add_pad_template (element_class,
1113 gst_pad_template_new ("sink", GST_PAD_SINK, GST_PAD_ALWAYS,
1115 gst_element_class_add_pad_template (element_class,
1116 gst_pad_template_new ("src", GST_PAD_SRC, GST_PAD_ALWAYS,
1119 gst_caps_unref (cdata->sink_caps);
1120 gst_caps_unref (cdata->src_caps);
1124 /* Probing functions */
1126 gst_v4l2_is_video_enc (GstCaps * sink_caps, GstCaps * src_caps,
1127 GstCaps * codec_caps)
1129 gboolean ret = FALSE;
1130 gboolean (*check_caps) (const GstCaps *, const GstCaps *);
1133 check_caps = gst_caps_can_intersect;
1135 codec_caps = gst_v4l2_object_get_codec_caps ();
1136 check_caps = gst_caps_is_subset;
1139 if (gst_caps_is_subset (sink_caps, gst_v4l2_object_get_raw_caps ())
1140 && check_caps (src_caps, codec_caps))
1147 gst_v4l2_video_enc_register (GstPlugin * plugin, GType type,
1148 const char *codec, const gchar * basename, const gchar * device_path,
1149 GstCaps * sink_caps, GstCaps * codec_caps, GstCaps * src_caps)
1151 GstCaps *filtered_caps;
1152 GTypeQuery type_query;
1153 GTypeInfo type_info = { 0, };
1156 GstV4l2VideoEncCData *cdata;
1158 filtered_caps = gst_caps_intersect (src_caps, codec_caps);
1160 cdata = g_new0 (GstV4l2VideoEncCData, 1);
1161 cdata->device = g_strdup (device_path);
1162 cdata->sink_caps = gst_caps_ref (sink_caps);
1163 cdata->src_caps = gst_caps_ref (filtered_caps);
1165 g_type_query (type, &type_query);
1166 memset (&type_info, 0, sizeof (type_info));
1167 type_info.class_size = type_query.class_size;
1168 type_info.instance_size = type_query.instance_size;
1169 type_info.class_init = gst_v4l2_video_enc_subclass_init;
1170 type_info.class_data = cdata;
1171 type_info.instance_init = gst_v4l2_video_enc_subinstance_init;
1173 /* The first encoder to be registered should use a constant name, like
1174 * v4l2h264enc, for any additional encoders, we create unique names. Encoder
1175 * names may change between boots, so this should help gain stable names for
1176 * the most common use cases. */
1177 type_name = g_strdup_printf ("v4l2%senc", codec);
1179 if (g_type_from_name (type_name) != 0) {
1181 type_name = g_strdup_printf ("v4l2%s%senc", basename, codec);
1184 subtype = g_type_register_static (type, type_name, &type_info, 0);
1186 if (!gst_element_register (plugin, type_name, GST_RANK_PRIMARY + 1, subtype))
1187 GST_WARNING ("Failed to register plugin '%s'", type_name);