diff options
author | Benjamin Gaignard <benjamin.gaignard@collabora.com> | 2022-07-12 08:56:07 -0400 |
---|---|---|
committer | Nicolas Dufresne <nicolas.dufresne@collabora.com> | 2022-07-22 10:21:02 -0400 |
commit | 0eb7028570f9064cf54a8de9b3f9671c748350a9 (patch) | |
tree | 84416c03ad83612c31279a478adbe56b7bdd4bbc /subprojects/gst-plugins-bad/sys/v4l2codecs | |
parent | 70941696ed13310310801b167f2bb7c3b0675cd0 (diff) |
v4l2codecs: Add v4l2 stateless H265 decoder
Add v4l2 stateless H265 decoder.
Co-authored-by: Nicolas Dufresne <nicolas.dufresne@collabora.com>
Co-authored-by: Sebastian Fricke <sebastian.fricke@collabora.com>
Part-of: <https://gitlab.freedesktop.org/gstreamer/gstreamer/-/merge_requests/1079>
Diffstat (limited to 'subprojects/gst-plugins-bad/sys/v4l2codecs')
4 files changed, 1686 insertions, 0 deletions
diff --git a/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.c b/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.c new file mode 100644 index 0000000000..459437209e --- /dev/null +++ b/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.c @@ -0,0 +1,1622 @@ +/* GStreamer + * Copyright (C) 2020 Nicolas Dufresne <nicolas.dufresne@collabora.com> + * Copyright (C) 2020 Safran Passenger Innovations LLC + * + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Library General Public + * License as published by the Free Software Foundation; either + * version 2 of the License, or (at your option) any later version. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Library General Public License for more details. + * + * You should have received a copy of the GNU Library General Public + * License along with this library; if not, write to the + * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, + * Boston, MA 02110-1301, USA. + */ + +#ifdef HAVE_CONFIG_H +#include <config.h> +#endif + +#include "gstv4l2codecallocator.h" +#include "gstv4l2codech265dec.h" +#include "gstv4l2codecpool.h" +#include "gstv4l2format.h" +#include "linux/v4l2-controls.h" + +#define KERNEL_VERSION(a,b,c) (((a) << 16) + ((b) << 8) + (c)) + +#define V4L2_MIN_KERNEL_VER_MAJOR 5 +#define V4L2_MIN_KERNEL_VER_MINOR 18 +#define V4L2_MIN_KERNEL_VERSION KERNEL_VERSION(V4L2_MIN_KERNEL_VER_MAJOR, V4L2_MIN_KERNEL_VER_MINOR, 0) + +GST_DEBUG_CATEGORY_STATIC (v4l2_h265dec_debug); +#define GST_CAT_DEFAULT v4l2_h265dec_debug + +enum +{ + PROP_0, + PROP_LAST = PROP_0 +}; + +static GstStaticPadTemplate sink_template = +GST_STATIC_PAD_TEMPLATE (GST_VIDEO_DECODER_SINK_NAME, + GST_PAD_SINK, GST_PAD_ALWAYS, + GST_STATIC_CAPS ("video/x-h265, " + "stream-format=(string) { hvc1, hev1, byte-stream }, " + "alignment=(string) au") + ); + +static GstStaticPadTemplate src_template = +GST_STATIC_PAD_TEMPLATE (GST_VIDEO_DECODER_SRC_NAME, + GST_PAD_SRC, GST_PAD_ALWAYS, + GST_STATIC_CAPS (GST_VIDEO_CAPS_MAKE (GST_V4L2_DEFAULT_VIDEO_FORMATS))); + +struct _GstV4l2CodecH265Dec +{ + GstH265Decoder parent; + GstV4l2Decoder *decoder; + GstVideoCodecState *output_state; + GstVideoInfo vinfo; + gint display_width; + gint display_height; + gint coded_width; + gint coded_height; + guint bitdepth; + guint chroma_format_idc; + guint num_slices; + gboolean first_slice; + + GstV4l2CodecAllocator *sink_allocator; + GstV4l2CodecAllocator *src_allocator; + GstV4l2CodecPool *src_pool; + gint min_pool_size; + gboolean has_videometa; + gboolean need_negotiation; + gboolean copy_frames; + gboolean need_sequence; + + struct v4l2_ctrl_hevc_sps sps; + struct v4l2_ctrl_hevc_pps pps; + struct v4l2_ctrl_hevc_scaling_matrix scaling_matrix; + struct v4l2_ctrl_hevc_decode_params decode_params; + GArray *slice_params; + GArray *entry_point_offsets; + + enum v4l2_stateless_hevc_decode_mode decode_mode; + enum v4l2_stateless_hevc_start_code start_code; + + GstMemory *bitstream; + GstMapInfo bitstream_map; + + gboolean support_scaling_matrix; + gboolean support_slice_parameters; + gboolean support_entry_point_offsets; + + GstVideoConverter *convert; + gboolean need_crop; + gint crop_rect_width, crop_rect_height; + gint crop_rect_x, crop_rect_y; +}; + +G_DEFINE_ABSTRACT_TYPE (GstV4l2CodecH265Dec, gst_v4l2_codec_h265_dec, + GST_TYPE_H265_DECODER); + +#define parent_class gst_v4l2_codec_h265_dec_parent_class + +static gboolean +is_frame_based (GstV4l2CodecH265Dec * self) +{ + return (self->decode_mode == + V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED) && + !self->support_slice_parameters; +} + +static gboolean +is_slice_based (GstV4l2CodecH265Dec * self) +{ + return self->decode_mode == V4L2_STATELESS_HEVC_DECODE_MODE_SLICE_BASED; +} + +static gboolean +is_frame_based_with_slices (GstV4l2CodecH265Dec * self) +{ + return (self->decode_mode == + V4L2_STATELESS_HEVC_DECODE_MODE_FRAME_BASED) && + self->support_slice_parameters; +} + +static gboolean +needs_start_codes (GstV4l2CodecH265Dec * self) +{ + return self->start_code == V4L2_STATELESS_HEVC_START_CODE_ANNEX_B; +} + + +static gboolean +gst_v4l2_codec_h265_dec_open (GstVideoDecoder * decoder) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + /* *INDENT-OFF* */ + struct v4l2_ext_control control[] = { + { + .id = V4L2_CID_STATELESS_HEVC_DECODE_MODE, + }, + { + .id = V4L2_CID_STATELESS_HEVC_START_CODE, + }, + }; + struct v4l2_ext_control scaling_matrix[] = { + { + .id = V4L2_CID_STATELESS_HEVC_SCALING_MATRIX, + .ptr = &self->scaling_matrix, + .size = sizeof (self->scaling_matrix), + }, + }; + /* *INDENT-ON* */ + + if (!gst_v4l2_decoder_open (self->decoder)) { + GST_ELEMENT_ERROR (self, RESOURCE, OPEN_READ_WRITE, + ("Failed to open H265 decoder"), + ("gst_v4l2_decoder_open() failed: %s", g_strerror (errno))); + return FALSE; + } + + if (!gst_v4l2_decoder_get_controls (self->decoder, control, + G_N_ELEMENTS (control))) { + GST_ELEMENT_ERROR (self, RESOURCE, OPEN_READ_WRITE, + ("Driver did not report framing and start code method."), + ("gst_v4l2_decoder_get_controls() failed: %s", g_strerror (errno))); + return FALSE; + } + + self->support_scaling_matrix = + gst_v4l2_decoder_get_controls (self->decoder, scaling_matrix, + G_N_ELEMENTS (scaling_matrix)); + + self->support_slice_parameters = + gst_v4l2_decoder_query_control_size (self->decoder, + V4L2_CID_STATELESS_HEVC_SLICE_PARAMS, NULL); + + self->support_entry_point_offsets = + gst_v4l2_decoder_query_control_size (self->decoder, + V4L2_CID_STATELESS_HEVC_ENTRY_POINT_OFFSETS, NULL); + + self->decode_mode = control[0].value; + self->start_code = control[1].value; + + GST_INFO_OBJECT (self, "Opened H265 %s decoder %s", + is_frame_based (self) ? "frame based" : is_slice_based (self) ? + "slice based" : "frame based with slices", + needs_start_codes (self) ? "using start-codes" : "without start-codes"); + gst_h265_decoder_set_process_ref_pic_lists (GST_H265_DECODER (self), + is_slice_based (self) || is_frame_based_with_slices (self)); + + return TRUE; +} + +static gboolean +gst_v4l2_codec_h265_dec_close (GstVideoDecoder * decoder) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + gst_v4l2_decoder_close (self->decoder); + return TRUE; +} + +static void +gst_v4l2_codec_h265_dec_reset_allocation (GstV4l2CodecH265Dec * self) +{ + if (self->sink_allocator) { + gst_v4l2_codec_allocator_detach (self->sink_allocator); + g_clear_object (&self->sink_allocator); + } + + if (self->src_allocator) { + gst_v4l2_codec_allocator_detach (self->src_allocator); + g_clear_object (&self->src_allocator); + g_clear_object (&self->src_pool); + } +} + +static gboolean +gst_v4l2_codec_h265_dec_stop (GstVideoDecoder * decoder) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + + gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SINK); + gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SRC); + + gst_v4l2_codec_h265_dec_reset_allocation (self); + + if (self->output_state) + gst_video_codec_state_unref (self->output_state); + self->output_state = NULL; + + return GST_VIDEO_DECODER_CLASS (parent_class)->stop (decoder); +} + +static gint +get_pixel_bitdepth (GstV4l2CodecH265Dec * self) +{ + gint depth; + + switch (self->chroma_format_idc) { + case 0: + /* 4:0:0 */ + depth = self->bitdepth; + break; + case 1: + /* 4:2:0 */ + depth = self->bitdepth + self->bitdepth / 2; + break; + case 2: + /* 4:2:2 */ + depth = 2 * self->bitdepth; + break; + case 3: + /* 4:4:4 */ + depth = 3 * self->bitdepth; + break; + default: + GST_WARNING_OBJECT (self, "Unsupported chroma format %i", + self->chroma_format_idc); + depth = 0; + break; + } + + return depth; +} + +static gboolean +gst_v4l2_codec_h265_dec_negotiate (GstVideoDecoder * decoder) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + GstH265Decoder *h265dec = GST_H265_DECODER (decoder); + /* *INDENT-OFF* */ + struct v4l2_ext_control control[] = { + { + .id = V4L2_CID_STATELESS_HEVC_SPS, + .ptr = &self->sps, + .size = sizeof (self->sps), + }, + }; + /* *INDENT-ON* */ + GstCaps *filter, *caps; + + /* Ignore downstream renegotiation request. */ + if (!self->need_negotiation) + return TRUE; + self->need_negotiation = FALSE; + + GST_DEBUG_OBJECT (self, "Negotiate"); + + gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SINK); + gst_v4l2_decoder_streamoff (self->decoder, GST_PAD_SRC); + + gst_v4l2_codec_h265_dec_reset_allocation (self); + + if (!gst_v4l2_decoder_set_sink_fmt (self->decoder, V4L2_PIX_FMT_HEVC_SLICE, + self->coded_width, self->coded_height, get_pixel_bitdepth (self))) { + GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, + ("Failed to configure H265 decoder"), + ("gst_v4l2_decoder_set_sink_fmt() failed: %s", g_strerror (errno))); + gst_v4l2_decoder_close (self->decoder); + return FALSE; + } + + if (!gst_v4l2_decoder_set_controls (self->decoder, NULL, control, + G_N_ELEMENTS (control))) { + GST_ELEMENT_ERROR (decoder, RESOURCE, WRITE, + ("Driver does not support the selected stream."), (NULL)); + return FALSE; + } + + filter = gst_v4l2_decoder_enum_src_formats (self->decoder); + if (!filter) { + GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, + ("No supported decoder output formats"), (NULL)); + return FALSE; + } + GST_DEBUG_OBJECT (self, "Supported output formats: %" GST_PTR_FORMAT, filter); + + caps = gst_pad_peer_query_caps (decoder->srcpad, filter); + gst_caps_unref (filter); + GST_DEBUG_OBJECT (self, "Peer supported formats: %" GST_PTR_FORMAT, caps); + + if (!gst_v4l2_decoder_select_src_format (self->decoder, caps, &self->vinfo)) { + GST_ELEMENT_ERROR (self, CORE, NEGOTIATION, + ("Unsupported bitdepth/chroma format"), + ("No support for %ux%u %ubit chroma IDC %i", self->coded_width, + self->coded_height, self->bitdepth, self->chroma_format_idc)); + gst_caps_unref (caps); + return FALSE; + } + gst_caps_unref (caps); + + if (self->output_state) + gst_video_codec_state_unref (self->output_state); + + self->output_state = + gst_video_decoder_set_output_state (GST_VIDEO_DECODER (self), + self->vinfo.finfo->format, self->display_width, + self->display_height, h265dec->input_state); + + self->output_state->caps = gst_video_info_to_caps (&self->output_state->info); + + if (GST_VIDEO_DECODER_CLASS (parent_class)->negotiate (decoder)) { + if (!gst_v4l2_decoder_streamon (self->decoder, GST_PAD_SINK)) { + GST_ELEMENT_ERROR (self, RESOURCE, FAILED, + ("Could not enable the decoder driver."), + ("VIDIOC_STREAMON(SINK) failed: %s", g_strerror (errno))); + return FALSE; + } + + if (!gst_v4l2_decoder_streamon (self->decoder, GST_PAD_SRC)) { + GST_ELEMENT_ERROR (self, RESOURCE, FAILED, + ("Could not enable the decoder driver."), + ("VIDIOC_STREAMON(SRC) failed: %s", g_strerror (errno))); + return FALSE; + } + + return TRUE; + } + + return FALSE; +} + +static gboolean +gst_v4l2_codec_h265_dec_decide_allocation (GstVideoDecoder * decoder, + GstQuery * query) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + guint min = 0; + + self->has_videometa = gst_query_find_allocation_meta (query, + GST_VIDEO_META_API_TYPE, NULL); + + g_clear_object (&self->src_pool); + g_clear_object (&self->src_allocator); + + if (gst_query_get_n_allocation_pools (query) > 0) + gst_query_parse_nth_allocation_pool (query, 0, NULL, NULL, &min, NULL); + + min = MAX (2, min); + + self->sink_allocator = gst_v4l2_codec_allocator_new (self->decoder, + GST_PAD_SINK, self->min_pool_size + 2); + self->src_allocator = gst_v4l2_codec_allocator_new (self->decoder, + GST_PAD_SRC, self->min_pool_size + min + 1); + self->src_pool = gst_v4l2_codec_pool_new (self->src_allocator, &self->vinfo); + + /* Our buffer pool is internal, we will let the base class create a video + * pool, and use it if we are running out of buffers or if downstream does + * not support GstVideoMeta */ + return GST_VIDEO_DECODER_CLASS (parent_class)->decide_allocation + (decoder, query); +} + +static void +gst_v4l2_codec_h265_dec_fill_sequence (GstV4l2CodecH265Dec * self, + const GstH265SPS * sps) +{ + /* Whenever we update teh sps, we need to send it again */ + self->need_sequence = TRUE; + + /* *INDENT-OFF* */ + self->sps = (struct v4l2_ctrl_hevc_sps) { + .video_parameter_set_id = sps->vps->id, + .seq_parameter_set_id = sps->id, + .pic_width_in_luma_samples = sps->pic_width_in_luma_samples, + .pic_height_in_luma_samples = sps->pic_height_in_luma_samples, + .bit_depth_luma_minus8 = sps->bit_depth_luma_minus8, + .bit_depth_chroma_minus8 = sps->bit_depth_chroma_minus8, + .log2_max_pic_order_cnt_lsb_minus4 = sps->log2_max_pic_order_cnt_lsb_minus4, + .sps_max_dec_pic_buffering_minus1 = sps->max_dec_pic_buffering_minus1[0], + .sps_max_num_reorder_pics = sps->max_num_reorder_pics[0], + .sps_max_latency_increase_plus1 = sps->max_latency_increase_plus1[0], + .log2_min_luma_coding_block_size_minus3 = sps->log2_min_luma_coding_block_size_minus3, + .log2_diff_max_min_luma_coding_block_size = sps->log2_diff_max_min_luma_coding_block_size, + .log2_min_luma_transform_block_size_minus2 = sps->log2_min_transform_block_size_minus2, + .log2_diff_max_min_luma_transform_block_size = sps->log2_diff_max_min_transform_block_size, + .max_transform_hierarchy_depth_inter = sps->max_transform_hierarchy_depth_inter, + .max_transform_hierarchy_depth_intra = sps->max_transform_hierarchy_depth_intra, + .num_short_term_ref_pic_sets = sps->num_short_term_ref_pic_sets, + .num_long_term_ref_pics_sps = sps->num_long_term_ref_pics_sps, + .chroma_format_idc = sps->chroma_format_idc, + .sps_max_sub_layers_minus1 = sps->max_sub_layers_minus1, + .flags = (sps->separate_colour_plane_flag ? V4L2_HEVC_SPS_FLAG_SEPARATE_COLOUR_PLANE : 0) | + (sps->scaling_list_enabled_flag ? V4L2_HEVC_SPS_FLAG_SCALING_LIST_ENABLED : 0) | + (sps->amp_enabled_flag ? V4L2_HEVC_SPS_FLAG_AMP_ENABLED : 0) | + (sps->sample_adaptive_offset_enabled_flag ? V4L2_HEVC_SPS_FLAG_SAMPLE_ADAPTIVE_OFFSET : 0) | + (sps->long_term_ref_pics_present_flag ? V4L2_HEVC_SPS_FLAG_LONG_TERM_REF_PICS_PRESENT : 0) | + (sps->temporal_mvp_enabled_flag ? V4L2_HEVC_SPS_FLAG_SPS_TEMPORAL_MVP_ENABLED : 0) | + (sps->strong_intra_smoothing_enabled_flag ? V4L2_HEVC_SPS_FLAG_STRONG_INTRA_SMOOTHING_ENABLED : 0), + }; + /* *INDENT-ON* */ + if (sps->pcm_enabled_flag) { + self->sps.pcm_sample_bit_depth_luma_minus1 = + sps->pcm_sample_bit_depth_luma_minus1; + self->sps.pcm_sample_bit_depth_chroma_minus1 = + sps->pcm_sample_bit_depth_chroma_minus1; + self->sps.log2_min_pcm_luma_coding_block_size_minus3 = + sps->log2_min_pcm_luma_coding_block_size_minus3; + self->sps.log2_diff_max_min_pcm_luma_coding_block_size = + sps->log2_diff_max_min_pcm_luma_coding_block_size; + self->sps.flags |= + V4L2_HEVC_SPS_FLAG_PCM_ENABLED | (sps->pcm_loop_filter_disabled_flag ? + V4L2_HEVC_SPS_FLAG_PCM_LOOP_FILTER_DISABLED : 0); + } +} + +static void +gst_v4l2_codec_h265_dec_fill_pps (GstV4l2CodecH265Dec * self, GstH265PPS * pps) +{ + gint i; + + /* *INDENT-OFF* */ + self->pps = (struct v4l2_ctrl_hevc_pps) { + .pic_parameter_set_id = pps->id, + .num_extra_slice_header_bits = pps->num_extra_slice_header_bits, + .num_ref_idx_l0_default_active_minus1 = pps->num_ref_idx_l0_default_active_minus1, + .num_ref_idx_l1_default_active_minus1 = pps->num_ref_idx_l1_default_active_minus1, + .init_qp_minus26 = pps->init_qp_minus26, + .diff_cu_qp_delta_depth = pps->diff_cu_qp_delta_depth, + .pps_cb_qp_offset = pps->cb_qp_offset, + .pps_cr_qp_offset = pps->cr_qp_offset, + .pps_beta_offset_div2 = pps->beta_offset_div2, + .pps_tc_offset_div2 = pps->tc_offset_div2, + .log2_parallel_merge_level_minus2 = pps->log2_parallel_merge_level_minus2, + .flags = (pps->dependent_slice_segments_enabled_flag ? V4L2_HEVC_PPS_FLAG_DEPENDENT_SLICE_SEGMENT_ENABLED : 0) | + (pps->output_flag_present_flag ? V4L2_HEVC_PPS_FLAG_OUTPUT_FLAG_PRESENT : 0) | + (pps->sign_data_hiding_enabled_flag ? V4L2_HEVC_PPS_FLAG_SIGN_DATA_HIDING_ENABLED : 0) | + (pps->cabac_init_present_flag ? V4L2_HEVC_PPS_FLAG_CABAC_INIT_PRESENT : 0) | + (pps->constrained_intra_pred_flag ? V4L2_HEVC_PPS_FLAG_CONSTRAINED_INTRA_PRED : 0) | + (pps->transform_skip_enabled_flag ? V4L2_HEVC_PPS_FLAG_TRANSFORM_SKIP_ENABLED : 0) | + (pps->cu_qp_delta_enabled_flag ? V4L2_HEVC_PPS_FLAG_CU_QP_DELTA_ENABLED : 0) | + (pps->slice_chroma_qp_offsets_present_flag ? V4L2_HEVC_PPS_FLAG_PPS_SLICE_CHROMA_QP_OFFSETS_PRESENT : 0) | + (pps->weighted_pred_flag ? V4L2_HEVC_PPS_FLAG_WEIGHTED_PRED : 0) | + (pps->weighted_bipred_flag ? V4L2_HEVC_PPS_FLAG_WEIGHTED_BIPRED : 0) | + (pps->transquant_bypass_enabled_flag ? V4L2_HEVC_PPS_FLAG_TRANSQUANT_BYPASS_ENABLED : 0) | + (pps->tiles_enabled_flag ? V4L2_HEVC_PPS_FLAG_TILES_ENABLED : 0) | + (pps->entropy_coding_sync_enabled_flag ? V4L2_HEVC_PPS_FLAG_ENTROPY_CODING_SYNC_ENABLED : 0) | + (pps->loop_filter_across_tiles_enabled_flag ? V4L2_HEVC_PPS_FLAG_LOOP_FILTER_ACROSS_TILES_ENABLED : 0) | + (pps->loop_filter_across_slices_enabled_flag ? V4L2_HEVC_PPS_FLAG_PPS_LOOP_FILTER_ACROSS_SLICES_ENABLED : 0 ) | + (pps->deblocking_filter_override_enabled_flag ? V4L2_HEVC_PPS_FLAG_DEBLOCKING_FILTER_OVERRIDE_ENABLED : 0) | + (pps->deblocking_filter_disabled_flag ? V4L2_HEVC_PPS_FLAG_PPS_DISABLE_DEBLOCKING_FILTER : 0) | + (pps->lists_modification_present_flag ? V4L2_HEVC_PPS_FLAG_LISTS_MODIFICATION_PRESENT : 0) | + (pps->slice_segment_header_extension_present_flag ? V4L2_HEVC_PPS_FLAG_SLICE_SEGMENT_HEADER_EXTENSION_PRESENT : 0) | + (pps->deblocking_filter_control_present_flag ? V4L2_HEVC_PPS_FLAG_DEBLOCKING_FILTER_CONTROL_PRESENT : 0) | + (pps->uniform_spacing_flag ? V4L2_HEVC_PPS_FLAG_UNIFORM_SPACING : 0), + }; + /* *INDENT-ON* */ + + if (pps->tiles_enabled_flag) { + self->pps.num_tile_columns_minus1 = pps->num_tile_columns_minus1; + self->pps.num_tile_rows_minus1 = pps->num_tile_rows_minus1; + + /* This should not be needed if we use uniform spacing, but Cedrus driver + * depends on it for now. As GStreamer parser do set that, lets just pass + * the information we have. */ + for (i = 0; i <= pps->num_tile_columns_minus1; i++) + self->pps.column_width_minus1[i] = pps->column_width_minus1[i]; + for (i = 0; i <= pps->num_tile_rows_minus1; i++) + self->pps.row_height_minus1[i] = pps->row_height_minus1[i]; + } +} + +static void +gst_v4l2_codec_h265_dec_fill_scaling_matrix (GstV4l2CodecH265Dec * self, + GstH265PPS * pps) +{ + GstH265ScalingList *sl; + gint i; + + if (!pps->sps->scaling_list_enabled_flag) { + memset (&self->scaling_matrix, 0, sizeof (self->scaling_matrix)); + return; + } + + if (pps->scaling_list_data_present_flag) + sl = &pps->scaling_list; + else if (pps->sps->scaling_list_data_present_flag) + sl = &pps->sps->scaling_list; + /* The default scaling list is strored in the pps */ + else + sl = &pps->scaling_list; + + for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_4x4); i++) + gst_h265_quant_matrix_4x4_get_raster_from_uprightdiagonal + (self->scaling_matrix.scaling_list_4x4[i], sl->scaling_lists_4x4[i]); + + for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_8x8); i++) + gst_h265_quant_matrix_8x8_get_raster_from_uprightdiagonal + (self->scaling_matrix.scaling_list_8x8[i], sl->scaling_lists_8x8[i]); + + for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_16x16); i++) + gst_h265_quant_matrix_16x16_get_raster_from_uprightdiagonal + (self->scaling_matrix.scaling_list_16x16[i], + sl->scaling_lists_16x16[i]); + + for (i = 0; i < G_N_ELEMENTS (sl->scaling_lists_32x32); i++) + gst_h265_quant_matrix_32x32_get_raster_from_uprightdiagonal + (self->scaling_matrix.scaling_list_32x32[i], + sl->scaling_lists_32x32[i]); + + for (i = 0; i < G_N_ELEMENTS (sl->scaling_list_dc_coef_minus8_16x16); i++) + self->scaling_matrix.scaling_list_dc_coef_16x16[i] = + sl->scaling_list_dc_coef_minus8_16x16[i] + 8; + + for (i = 0; i < G_N_ELEMENTS (sl->scaling_list_dc_coef_minus8_32x32); i++) + self->scaling_matrix.scaling_list_dc_coef_32x32[i] = + sl->scaling_list_dc_coef_minus8_32x32[i] + 8; +} + +static guint +get_slice_header_byte_offset (GstH265Slice * slice) +{ + guint epb_count, nal_header_bytes; + + epb_count = slice->header.n_emulation_prevention_bytes; + nal_header_bytes = slice->nalu.header_bytes; + + return nal_header_bytes + (slice->header.header_size + 7) / 8 - epb_count; +} + +static void +gst_v4l2_codec_h265_dec_fill_slice_params (GstV4l2CodecH265Dec * self, + GstH265Slice * slice, GstH265Picture * picture) +{ + gint n = self->num_slices++; + GstH265SliceHdr *slice_hdr = &slice->header; + GstH265PPS *pps = slice_hdr->pps; + gsize slice_size = slice->nalu.size; + gsize sc_offset = 0; + struct v4l2_ctrl_hevc_slice_params *params; + gint i, j; + gint chroma_weight, chroma_log2_weight_denom; + /* TODO adjust this if sps_ext is later supported */ + const gint32 WpOffsetHalfRangeC = 1 << 7; + + /* Ensure array is large enough */ + if (self->slice_params->len < self->num_slices) + g_array_set_size (self->slice_params, self->slice_params->len * 2); + + if (needs_start_codes (self)) + sc_offset = 3; + + /* *INDENT-OFF* */ + params = &g_array_index (self->slice_params, struct v4l2_ctrl_hevc_slice_params, n); + *params = (struct v4l2_ctrl_hevc_slice_params) { + .bit_size = (slice_size + sc_offset) * 8, + .data_byte_offset = get_slice_header_byte_offset (slice) + sc_offset, + .num_entry_point_offsets = slice_hdr->num_entry_point_offsets, + .nal_unit_type = slice->nalu.type, + .nuh_temporal_id_plus1 = slice->nalu.temporal_id_plus1, + .slice_type = slice_hdr->type, + .colour_plane_id = slice_hdr->colour_plane_id, + .slice_pic_order_cnt = picture->pic_order_cnt, + .num_ref_idx_l0_active_minus1 = slice_hdr->num_ref_idx_l0_active_minus1, + .num_ref_idx_l1_active_minus1 = slice_hdr->num_ref_idx_l1_active_minus1, + .collocated_ref_idx = slice_hdr->collocated_ref_idx, + .five_minus_max_num_merge_cand = slice_hdr->five_minus_max_num_merge_cand, + .slice_qp_delta = slice_hdr->qp_delta, + .slice_cb_qp_offset = slice_hdr->cb_qp_offset, + .slice_cr_qp_offset = slice_hdr->cr_qp_offset, + .slice_act_y_qp_offset = slice_hdr->slice_act_y_qp_offset, + .slice_act_cb_qp_offset = slice_hdr->slice_act_cb_qp_offset, + .slice_act_cr_qp_offset = slice_hdr->slice_act_cr_qp_offset, + .slice_beta_offset_div2 = slice_hdr->beta_offset_div2, + .slice_tc_offset_div2 = slice_hdr->tc_offset_div2, + .pic_struct = picture->pic_struct, + .slice_segment_addr = slice_hdr->segment_address, + .short_term_ref_pic_set_size = slice_hdr->short_term_ref_pic_set_size, + .long_term_ref_pic_set_size = slice_hdr->long_term_ref_pic_set_size, + .pred_weight_table = (struct v4l2_hevc_pred_weight_table) { + .luma_log2_weight_denom = slice_hdr->pred_weight_table.luma_log2_weight_denom, + .delta_chroma_log2_weight_denom = slice_hdr->pred_weight_table.delta_chroma_log2_weight_denom, + }, + .flags = + (slice_hdr->sao_luma_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_SAO_LUMA : 0) | + (slice_hdr->sao_chroma_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_SAO_CHROMA : 0) | + (slice_hdr->temporal_mvp_enabled_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_TEMPORAL_MVP_ENABLED : 0) | + (slice_hdr->mvd_l1_zero_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_MVD_L1_ZERO : 0) | + (slice_hdr->cabac_init_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_CABAC_INIT : 0) | + (slice_hdr->collocated_from_l0_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_COLLOCATED_FROM_L0 : 0) | + (slice_hdr->use_integer_mv_flag ? V4L2_HEVC_SLICE_PARAMS_FLAG_USE_INTEGER_MV : 0) | + (slice_hdr->deblocking_filter_disabled_flag ? + V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_DEBLOCKING_FILTER_DISABLED : 0) | + (slice_hdr->loop_filter_across_slices_enabled_flag ? + V4L2_HEVC_SLICE_PARAMS_FLAG_SLICE_LOOP_FILTER_ACROSS_SLICES_ENABLED : 0) | + (slice_hdr->dependent_slice_segment_flag ? + V4L2_HEVC_SLICE_PARAMS_FLAG_DEPENDENT_SLICE_SEGMENT : 0), + }; + /* *INDENT-ON* */ + + for (i = 0; i < slice_hdr->num_entry_point_offsets; i++) { + guint32 entry_point_offset = slice_hdr->entry_point_offset_minus1[i] + 1; + g_array_append_val (self->entry_point_offsets, entry_point_offset); + } + + if (GST_H265_IS_I_SLICE (slice_hdr) || + (!pps->weighted_pred_flag && GST_H265_IS_P_SLICE (slice_hdr)) || + (!pps->weighted_bipred_flag && GST_H265_IS_B_SLICE (slice_hdr))) + return; + + for (i = 0; i <= slice_hdr->num_ref_idx_l0_active_minus1; i++) { + if (!slice_hdr->pred_weight_table.luma_weight_l0_flag[i]) + continue; + + params->pred_weight_table.delta_luma_weight_l0[i] = + slice_hdr->pred_weight_table.delta_luma_weight_l0[i]; + params->pred_weight_table.luma_offset_l0[i] = + slice_hdr->pred_weight_table.luma_offset_l0[i]; + } + + chroma_log2_weight_denom = + slice_hdr->pred_weight_table.luma_log2_weight_denom + + slice_hdr->pred_weight_table.delta_chroma_log2_weight_denom; + + if (slice_hdr->pps->sps->chroma_array_type != 0) { + + for (i = 0; i <= slice_hdr->num_ref_idx_l0_active_minus1; i++) { + if (!slice_hdr->pred_weight_table.chroma_weight_l0_flag[i]) + continue; + + for (j = 0; j < 2; j++) { + gint16 delta_chroma_offset_l0 = + slice_hdr->pred_weight_table.delta_chroma_offset_l0[i][j]; + gint chroma_offset; + + params->pred_weight_table.delta_chroma_weight_l0[i][j] = + slice_hdr->pred_weight_table.delta_chroma_weight_l0[i][j]; + + /* Find ChromaWeightL0 */ + chroma_weight = (1 << chroma_log2_weight_denom) + + slice_hdr->pred_weight_table.delta_chroma_weight_l0[i][j]; + chroma_offset = WpOffsetHalfRangeC + delta_chroma_offset_l0 - + ((WpOffsetHalfRangeC * chroma_weight) >> chroma_log2_weight_denom); + + /* 7-56 */ + params->pred_weight_table.chroma_offset_l0[i][j] = + CLAMP (chroma_offset, -WpOffsetHalfRangeC, WpOffsetHalfRangeC - 1); + } + } + } + + /* Skip l1 if this is not a B-Frame. */ + if (!GST_H265_IS_B_SLICE (slice_hdr)) + return; + + for (i = 0; i <= slice_hdr->num_ref_idx_l1_active_minus1; i++) { + if (!slice_hdr->pred_weight_table.luma_weight_l1_flag[i]) + continue; + + params->pred_weight_table.delta_luma_weight_l1[i] = + slice_hdr->pred_weight_table.delta_luma_weight_l1[i]; + params->pred_weight_table.luma_offset_l1[i] = + slice_hdr->pred_weight_table.luma_offset_l1[i]; + } + + if (slice_hdr->pps->sps->chroma_array_type != 0) { + for (i = 0; i <= slice_hdr->num_ref_idx_l1_active_minus1; i++) { + if (!slice_hdr->pred_weight_table.chroma_weight_l1_flag[i]) + continue; + + for (j = 0; j < 2; j++) { + gint16 delta_chroma_offset_l1 = + slice_hdr->pred_weight_table.delta_chroma_offset_l1[i][j]; + gint chroma_offset; + + params->pred_weight_table.delta_chroma_weight_l1[i][j] = + slice_hdr->pred_weight_table.delta_chroma_weight_l1[i][j]; + + /* Find ChromaWeightL1 */ + chroma_weight = (1 << chroma_log2_weight_denom) + + slice_hdr->pred_weight_table.delta_chroma_weight_l1[i][j]; + + chroma_offset = WpOffsetHalfRangeC + delta_chroma_offset_l1 - + ((WpOffsetHalfRangeC * chroma_weight) >> chroma_log2_weight_denom); + + /* 7-56 */ + params->pred_weight_table.chroma_offset_l1[i][j] = + CLAMP (chroma_offset, -WpOffsetHalfRangeC, WpOffsetHalfRangeC - 1); + } + } + } +} + +static guint8 +lookup_dpb_index (struct v4l2_hevc_dpb_entry dpb[16], GstH265Picture * ref_pic) +{ + guint64 ref_ts; + gint i; + + /* Reference list may have wholes in case a ref is missing, we should mark + * the whole and avoid moving items in the list */ + if (!ref_pic) + return 0xff; + + ref_ts = (guint64) ref_pic->system_frame_number * 1000; + for (i = 0; i < 16; i++) { + if (dpb[i].timestamp == ref_ts) + return i; + } + + return 0xff; +} + +static void +gst_v4l2_codec_h265_dec_fill_references (GstV4l2CodecH265Dec * self, + GArray * ref_pic_list0, GArray * ref_pic_list1) +{ + struct v4l2_ctrl_hevc_slice_params *slice_params; + struct v4l2_ctrl_hevc_decode_params *decode_params = &self->decode_params; + gint i; + + slice_params = &g_array_index (self->slice_params, + struct v4l2_ctrl_hevc_slice_params, self->num_slices - 1); + + memset (slice_params->ref_idx_l0, 0xff, sizeof (slice_params->ref_idx_l0)); + memset (slice_params->ref_idx_l1, 0xff, sizeof (slice_params->ref_idx_l1)); + + for (i = 0; i < ref_pic_list0->len; i++) { + GstH265Picture *ref_pic = + g_array_index (ref_pic_list0, GstH265Picture *, i); + slice_params->ref_idx_l0[i] = + lookup_dpb_index (decode_params->dpb, ref_pic); + } + + for (i = 0; i < ref_pic_list1->len; i++) { + GstH265Picture *ref_pic = + g_array_index (ref_pic_list1, GstH265Picture *, i); + slice_params->ref_idx_l1[i] = + lookup_dpb_index (decode_params->dpb, ref_pic); + } +} + +static GstFlowReturn +gst_v4l2_codec_h265_dec_new_sequence (GstH265Decoder * decoder, + const GstH265SPS * sps, gint max_dpb_size) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + gint crop_width = sps->width; + gint crop_height = sps->height; + gboolean negotiation_needed = FALSE; + + if (self->vinfo.finfo->format == GST_VIDEO_FORMAT_UNKNOWN) + negotiation_needed = TRUE; + + /* TODO check if CREATE_BUFS is supported, and simply grow the pool */ + if (self->min_pool_size < max_dpb_size) { + self->min_pool_size = max_dpb_size; + negotiation_needed = TRUE; + } + + self->need_crop = FALSE; + if (sps->conformance_window_flag) { + crop_width = sps->crop_rect_width; + crop_height = sps->crop_rect_height; + self->crop_rect_width = sps->crop_rect_width; + self->crop_rect_height = sps->crop_rect_height; + self->crop_rect_x = sps->crop_rect_x; + self->crop_rect_y = sps->crop_rect_y; + + /* conformance_window_flag could be set but with zeroed + * parameters so check if we really need to crop */ + self->need_crop |= self->crop_rect_width != sps->width; + self->need_crop |= self->crop_rect_height != sps->height; + self->need_crop |= self->crop_rect_x != 0; + self->need_crop |= self->crop_rect_y != 0; + } + + /* TODO Check if current buffers are large enough, and reuse them */ + if (self->display_width != crop_width || self->display_height != crop_height + || self->coded_width != sps->width || self->coded_height != sps->height) { + self->display_width = crop_width; + self->display_height = crop_height; + self->coded_width = sps->width; + self->coded_height = sps->height; + negotiation_needed = TRUE; + GST_INFO_OBJECT (self, "Resolution changed to %dx%d (%ix%i)", + self->display_width, self->display_height, + self->coded_width, self->coded_height); + } + + if (self->bitdepth != sps->bit_depth_luma_minus8 + 8) { + self->bitdepth = sps->bit_depth_luma_minus8 + 8; + negotiation_needed = TRUE; + GST_INFO_OBJECT (self, "Bitdepth changed to %u", self->bitdepth); + } + + if (self->chroma_format_idc != sps->chroma_format_idc) { + self->chroma_format_idc = sps->chroma_format_idc; + negotiation_needed = TRUE; + GST_INFO_OBJECT (self, "Chroma format changed to %i", + self->chroma_format_idc); + } + + gst_v4l2_codec_h265_dec_fill_sequence (self, sps); + + if (negotiation_needed) { + self->need_negotiation = TRUE; + if (!gst_video_decoder_negotiate (GST_VIDEO_DECODER (self))) { + GST_ERROR_OBJECT (self, "Failed to negotiate with downstream"); + return GST_FLOW_NOT_NEGOTIATED; + } + } + + /* Check if we can zero-copy buffers */ + if (!self->has_videometa) { + GstVideoInfo ref_vinfo; + gint i; + + gst_video_info_set_format (&ref_vinfo, GST_VIDEO_INFO_FORMAT (&self->vinfo), + self->display_width, self->display_height); + + for (i = 0; i < GST_VIDEO_INFO_N_PLANES (&self->vinfo); i++) { + if (self->vinfo.stride[i] != ref_vinfo.stride[i] || + self->vinfo.offset[i] != ref_vinfo.offset[i]) { + GST_WARNING_OBJECT (self, + "GstVideoMeta support required, copying frames."); + self->copy_frames = TRUE; + break; + } + } + } else { + self->copy_frames = self->need_crop; + } + + return GST_FLOW_OK; +} + +static gboolean +gst_v4l2_codec_h265_dec_ensure_bitstream (GstV4l2CodecH265Dec * self) +{ + if (self->bitstream) + goto done; + + self->bitstream = gst_v4l2_codec_allocator_alloc (self->sink_allocator); + + if (!self->bitstream) { + GST_ELEMENT_ERROR (self, RESOURCE, NO_SPACE_LEFT, + ("Not enough memory to decode H265 stream."), (NULL)); + return FALSE; + } + + if (!gst_memory_map (self->bitstream, &self->bitstream_map, GST_MAP_WRITE)) { + GST_ELEMENT_ERROR (self, RESOURCE, WRITE, + ("Could not access bitstream memory for writing"), (NULL)); + g_clear_pointer (&self->bitstream, gst_memory_unref); + return FALSE; + } + +done: + /* We use this field to track how much we have written */ + self->bitstream_map.size = 0; + + return TRUE; +} + +static void +gst_v4l2_codec_h265_dec_fill_decode_params (GstV4l2CodecH265Dec * self, + GstH265Picture * picture, GstH265Slice * slice, GstH265Dpb * dpb) +{ + GstH265Decoder *decoder = (GstH265Decoder *) self; + GArray *refs = gst_h265_dpb_get_pictures_all (dpb); + gint i; + + /* *INDENT-OFF* */ + self->decode_params = (struct v4l2_ctrl_hevc_decode_params) { + .pic_order_cnt_val = picture->pic_order_cnt, + .num_poc_st_curr_before = decoder->NumPocStCurrBefore, + .num_poc_st_curr_after = decoder->NumPocStCurrAfter, + .num_poc_lt_curr = decoder->NumPocLtCurr, + .flags = + (GST_H265_IS_NAL_TYPE_IRAP (slice->nalu.type) ? V4L2_HEVC_DECODE_PARAM_FLAG_IRAP_PIC : 0) | + (GST_H265_IS_NAL_TYPE_IDR (slice->nalu.type) ? V4L2_HEVC_DECODE_PARAM_FLAG_IDR_PIC : 0) | + (slice->header.no_output_of_prior_pics_flag ? V4L2_HEVC_DECODE_PARAM_FLAG_NO_OUTPUT_OF_PRIOR : 0), + }; + /* *INDENT-ON* */ + + for (i = 0; i < refs->len; i++) { + GstH265Picture *ref_pic = g_array_index (refs, GstH265Picture *, i); + + if (!ref_pic->ref) + continue; + + /* *INDENT-OFF* */ + self->decode_params.dpb[self->decode_params.num_active_dpb_entries++] = + (struct v4l2_hevc_dpb_entry) { + /* + * The reference is multiplied by 1000 because it's wassed as micro + * seconds and this TS is nanosecond. + */ + .timestamp = (guint64) ref_pic->system_frame_number * 1000, + .flags = ref_pic->long_term ? V4L2_HEVC_DPB_ENTRY_LONG_TERM_REFERENCE : 0, + .field_pic = ref_pic->pic_struct, + .pic_order_cnt_val = ref_pic->pic_order_cnt, + }; + /* *INDENT-ON* */ + } + + for (i = 0; i < 16; i++) { + self->decode_params.poc_st_curr_before[i] = + lookup_dpb_index (self->decode_params.dpb, + decoder->RefPicSetStCurrBefore[i]); + self->decode_params.poc_st_curr_after[i] = + lookup_dpb_index (self->decode_params.dpb, + decoder->RefPicSetStCurrAfter[i]); + self->decode_params.poc_lt_curr[i] = + lookup_dpb_index (self->decode_params.dpb, decoder->RefPicSetLtCurr[i]); + } + + g_array_unref (refs); +} + +static GstFlowReturn +gst_v4l2_codec_h265_dec_start_picture (GstH265Decoder * decoder, + GstH265Picture * picture, GstH265Slice * slice, GstH265Dpb * dpb) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + + /* FIXME base class should not call us if negotiation failed */ + if (!self->sink_allocator) + return GST_FLOW_NOT_NEGOTIATED; + + if (!gst_v4l2_codec_h265_dec_ensure_bitstream (self)) + return GST_FLOW_ERROR; + + /* The base class will only emit new_sequence for allocation related changes + * in the SPS, make sure to keep the SPS upt-to-date */ + if (slice->header.pps->sps->id != self->sps.seq_parameter_set_id) + gst_v4l2_codec_h265_dec_fill_sequence (self, slice->header.pps->sps); + + gst_v4l2_codec_h265_dec_fill_pps (self, slice->header.pps); + gst_v4l2_codec_h265_dec_fill_scaling_matrix (self, slice->header.pps); + gst_v4l2_codec_h265_dec_fill_decode_params (self, picture, slice, dpb); + + self->first_slice = TRUE; + self->num_slices = 0; + g_array_set_size (self->entry_point_offsets, 0); + + return GST_FLOW_OK; +} + +static gboolean +gst_v4l2_codec_h265_dec_crop_output_buffer (GstV4l2CodecH265Dec * self, + GstVideoFrame * dest_frame, GstVideoFrame * src_frame) +{ + GstVideoInfo dst_info = dest_frame->info; + + dst_info.fps_n = src_frame->info.fps_n; + dst_info.fps_d = src_frame->info.fps_d; + + if (self->convert) { + gboolean new_convert = FALSE; + gint x = 0, y = 0, width = 0, height = 0; + const GstStructure *config = gst_video_converter_get_config (self->convert); + + if (!gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_X, &x) + || !gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_Y, &y) + || !gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_WIDTH, + &width) + || !gst_structure_get_int (config, GST_VIDEO_CONVERTER_OPT_SRC_HEIGHT, + &height)) + new_convert = TRUE; + + new_convert |= (self->crop_rect_x != x); + new_convert |= (self->crop_rect_y != y); + new_convert |= (self->crop_rect_width != width); + new_convert |= (self->crop_rect_height != height); + + /* No need to check dest, it always has (0,0) -> (width, height) */ + + if (new_convert) + g_clear_pointer (&self->convert, gst_video_converter_free); + } + + if (!self->convert) { + self->convert = gst_video_converter_new (&src_frame->info, &dst_info, + gst_structure_new ("options", + GST_VIDEO_CONVERTER_OPT_DITHER_METHOD, + GST_TYPE_VIDEO_DITHER_METHOD, GST_VIDEO_DITHER_NONE, + GST_VIDEO_CONVERTER_OPT_DITHER_QUANTIZATION, + G_TYPE_UINT, 0, + GST_VIDEO_CONVERTER_OPT_CHROMA_MODE, + GST_TYPE_VIDEO_CHROMA_MODE, GST_VIDEO_CHROMA_MODE_NONE, + GST_VIDEO_CONVERTER_OPT_MATRIX_MODE, + GST_TYPE_VIDEO_MATRIX_MODE, GST_VIDEO_MATRIX_MODE_NONE, + GST_VIDEO_CONVERTER_OPT_SRC_X, G_TYPE_INT, self->crop_rect_x, + GST_VIDEO_CONVERTER_OPT_SRC_Y, G_TYPE_INT, self->crop_rect_y, + GST_VIDEO_CONVERTER_OPT_SRC_WIDTH, G_TYPE_INT, + self->crop_rect_width, GST_VIDEO_CONVERTER_OPT_SRC_HEIGHT, + G_TYPE_INT, self->crop_rect_height, GST_VIDEO_CONVERTER_OPT_DEST_X, + G_TYPE_INT, 0, GST_VIDEO_CONVERTER_OPT_DEST_Y, G_TYPE_INT, 0, + GST_VIDEO_CONVERTER_OPT_DEST_WIDTH, G_TYPE_INT, self->display_width, + GST_VIDEO_CONVERTER_OPT_DEST_HEIGHT, G_TYPE_INT, + self->display_height, NULL)); + + if (!self->convert) { + GST_WARNING_OBJECT (self, "failed to create a video convert"); + return FALSE; + } + } + + gst_video_converter_frame (self->convert, src_frame, dest_frame); + + return TRUE; +} + +static gboolean +gst_v4l2_codec_h265_dec_copy_output_buffer (GstV4l2CodecH265Dec * self, + GstVideoCodecFrame * codec_frame) +{ + GstVideoFrame src_frame; + GstVideoFrame dest_frame; + GstVideoInfo dest_vinfo; + GstBuffer *buffer; + + gst_video_info_set_format (&dest_vinfo, GST_VIDEO_INFO_FORMAT (&self->vinfo), + self->display_width, self->display_height); + + buffer = gst_video_decoder_allocate_output_buffer (GST_VIDEO_DECODER (self)); + if (!buffer) + goto fail; + + if (!gst_video_frame_map (&src_frame, &self->vinfo, + codec_frame->output_buffer, GST_MAP_READ)) + goto fail; + + if (!gst_video_frame_map (&dest_frame, &dest_vinfo, buffer, GST_MAP_WRITE)) { + gst_video_frame_unmap (&dest_frame); + goto fail; + } + + if (self->need_crop) { + if (!gst_v4l2_codec_h265_dec_crop_output_buffer (self, &dest_frame, + &src_frame)) { + gst_video_frame_unmap (&src_frame); + gst_video_frame_unmap (&dest_frame); + GST_ERROR_OBJECT (self, "fail to apply the video crop."); + goto fail; + } + } else { + /* gst_video_frame_copy can crop this, but does not know, so let make it + * think it's all right */ + GST_VIDEO_INFO_WIDTH (&src_frame.info) = self->display_width; + GST_VIDEO_INFO_HEIGHT (&src_frame.info) = self->display_height; + + if (!gst_video_frame_copy (&dest_frame, &src_frame)) { + gst_video_frame_unmap (&src_frame); + gst_video_frame_unmap (&dest_frame); + goto fail; + } + } + + gst_video_frame_unmap (&src_frame); + gst_video_frame_unmap (&dest_frame); + gst_buffer_replace (&codec_frame->output_buffer, buffer); + gst_buffer_unref (buffer); + + return TRUE; + +fail: + GST_ERROR_OBJECT (self, "Failed copy output buffer."); + return FALSE; +} + +static GstFlowReturn +gst_v4l2_codec_h265_dec_output_picture (GstH265Decoder * decoder, + GstVideoCodecFrame * frame, GstH265Picture * picture) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + GstV4l2Request *request = gst_h265_picture_get_user_data (picture); + gint ret; + + GST_DEBUG_OBJECT (self, "Output picture %u", picture->system_frame_number); + + ret = gst_v4l2_request_set_done (request); + if (ret == 0) { + GST_ELEMENT_ERROR (self, STREAM, DECODE, + ("Decoding frame %u took too long", picture->system_frame_number), + (NULL)); + goto error; + } else if (ret < 0) { + GST_ELEMENT_ERROR (self, STREAM, DECODE, + ("Decoding request failed: %s", g_strerror (errno)), (NULL)); + goto error; + } + g_return_val_if_fail (frame->output_buffer, GST_FLOW_ERROR); + + if (gst_v4l2_request_failed (request)) { + GST_ELEMENT_ERROR (self, STREAM, DECODE, + ("Failed to decode frame %u", picture->system_frame_number), (NULL)); + goto error; + } + + /* Hold on reference buffers for the rest of the picture lifetime */ + gst_h265_picture_set_user_data (picture, + gst_buffer_ref (frame->output_buffer), (GDestroyNotify) gst_buffer_unref); + + if (self->copy_frames) + gst_v4l2_codec_h265_dec_copy_output_buffer (self, frame); + + gst_h265_picture_unref (picture); + + return gst_video_decoder_finish_frame (GST_VIDEO_DECODER (self), frame); + +error: + gst_video_decoder_drop_frame (GST_VIDEO_DECODER (self), frame); + gst_h265_picture_unref (picture); + + return GST_FLOW_ERROR; +} + +static void +gst_v4l2_codec_h265_dec_reset_picture (GstV4l2CodecH265Dec * self) +{ + if (self->bitstream) { + if (self->bitstream_map.memory) + gst_memory_unmap (self->bitstream, &self->bitstream_map); + g_clear_pointer (&self->bitstream, gst_memory_unref); + self->bitstream_map = (GstMapInfo) GST_MAP_INFO_INIT; + } + + self->num_slices = 0; + g_array_set_size (self->entry_point_offsets, 0); +} + +static gboolean +gst_v4l2_codec_h265_dec_ensure_output_buffer (GstV4l2CodecH265Dec * self, + GstVideoCodecFrame * frame) +{ + GstBuffer *buffer; + GstFlowReturn flow_ret; + + if (frame->output_buffer) + return TRUE; + + flow_ret = gst_buffer_pool_acquire_buffer (GST_BUFFER_POOL (self->src_pool), + &buffer, NULL); + if (flow_ret != GST_FLOW_OK) { + if (flow_ret == GST_FLOW_FLUSHING) + GST_DEBUG_OBJECT (self, "Frame decoding aborted, we are flushing."); + else + GST_ELEMENT_ERROR (self, RESOURCE, WRITE, + ("No more picture buffer available."), (NULL)); + return FALSE; + } + + frame->output_buffer = buffer; + return TRUE; +} + +static gboolean +gst_v4l2_codec_h265_dec_submit_bitstream (GstV4l2CodecH265Dec * self, + GstH265Picture * picture, guint flags) +{ + GstV4l2Request *prev_request, *request = NULL; + gsize bytesused; + gboolean ret = FALSE; + gint num_controls = 0; + + /* *INDENT-OFF* */ + /* Reserve space for controls */ + struct v4l2_ext_control control[] = { + { }, /* SPS */ + { }, /* PPS */ + { }, /* DECODE_PARAMS */ + { }, /* SLICE_PARAMS */ + { }, /* SCALING_MATRIX */ + { }, /* ENTRY_POINT_OFFSETS */ + }; + /* *INDENT-ON* */ + + prev_request = gst_h265_picture_get_user_data (picture); + + bytesused = self->bitstream_map.size; + gst_memory_unmap (self->bitstream, &self->bitstream_map); + self->bitstream_map = (GstMapInfo) GST_MAP_INFO_INIT; + gst_memory_resize (self->bitstream, 0, bytesused); + + if (prev_request) { + request = gst_v4l2_decoder_alloc_sub_request (self->decoder, prev_request, + self->bitstream); + } else { + GstVideoCodecFrame *frame; + + frame = gst_video_decoder_get_frame (GST_VIDEO_DECODER (self), + picture->system_frame_number); + g_return_val_if_fail (frame, FALSE); + + if (!gst_v4l2_codec_h265_dec_ensure_output_buffer (self, frame)) + goto done; + + request = gst_v4l2_decoder_alloc_request (self->decoder, + picture->system_frame_number, self->bitstream, frame->output_buffer); + + gst_video_codec_frame_unref (frame); + } + + if (!request) { + GST_ELEMENT_ERROR (self, RESOURCE, NO_SPACE_LEFT, + ("Failed to allocate a media request object."), (NULL)); + goto done; + } + + if (self->need_sequence) { + control[num_controls].id = V4L2_CID_STATELESS_HEVC_SPS; + control[num_controls].ptr = &self->sps; + control[num_controls].size = sizeof (self->sps); + num_controls++; + self->need_sequence = FALSE; + } + + if (self->first_slice) { + control[num_controls].id = V4L2_CID_STATELESS_HEVC_PPS; + control[num_controls].ptr = &self->pps; + control[num_controls].size = sizeof (self->pps); + num_controls++; + + if (self->support_scaling_matrix) { + control[num_controls].id = V4L2_CID_STATELESS_HEVC_SCALING_MATRIX; + control[num_controls].ptr = &self->scaling_matrix; + control[num_controls].size = sizeof (self->scaling_matrix); + num_controls++; + } + + control[num_controls].id = V4L2_CID_STATELESS_HEVC_DECODE_PARAMS; + control[num_controls].ptr = &self->decode_params; + control[num_controls].size = sizeof (self->decode_params); + num_controls++; + + self->first_slice = FALSE; + } + + /* Slice parameters are only filled for slice based or frame based with + * slices decoders */ + if (self->num_slices && !is_frame_based (self)) { + control[num_controls].id = V4L2_CID_STATELESS_HEVC_SLICE_PARAMS; + control[num_controls].ptr = self->slice_params->data; + control[num_controls].size = g_array_get_element_size (self->slice_params) + * self->num_slices; + num_controls++; + + if (self->support_entry_point_offsets && self->entry_point_offsets->len) { + control[num_controls].id = V4L2_CID_STATELESS_HEVC_ENTRY_POINT_OFFSETS; + control[num_controls].ptr = self->entry_point_offsets->data; + control[num_controls].size = + g_array_get_element_size (self->entry_point_offsets) + * self->entry_point_offsets->len; + num_controls++; + } + } + + if (num_controls > G_N_ELEMENTS (control)) + g_error ("Set too many controls, increase control[] size"); + + if (!gst_v4l2_decoder_set_controls (self->decoder, request, control, + num_controls)) { + GST_ELEMENT_ERROR (self, RESOURCE, WRITE, + ("Driver did not accept the bitstream parameters."), (NULL)); + goto done; + } + + if (!gst_v4l2_request_queue (request, flags)) { + GST_ELEMENT_ERROR (self, RESOURCE, WRITE, + ("Driver did not accept the decode request."), (NULL)); + goto done; + } + + gst_h265_picture_set_user_data (picture, g_steal_pointer (&request), + (GDestroyNotify) gst_v4l2_request_unref); + ret = TRUE; + +done: + if (request) + gst_v4l2_request_unref (request); + + gst_v4l2_codec_h265_dec_reset_picture (self); + + return ret; +} + +static GstFlowReturn +gst_v4l2_codec_h265_dec_decode_slice (GstH265Decoder * decoder, + GstH265Picture * picture, GstH265Slice * slice, + GArray * ref_pic_list0, GArray * ref_pic_list1) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + gsize sc_off = 0; + gsize nal_size; + guint8 *bitstream_data; + + if (!is_frame_based (self)) { + if (is_slice_based (self)) { + if (self->bitstream_map.size) { + /* In slice mode, we submit the pending slice asking the accelerator to + * hold on the picture */ + if (!gst_v4l2_codec_h265_dec_submit_bitstream (self, picture, + V4L2_BUF_FLAG_M2M_HOLD_CAPTURE_BUF) + || !gst_v4l2_codec_h265_dec_ensure_bitstream (self)) + return GST_FLOW_ERROR; + } + } + /* in frame based mode with slices, we need to provide the required data for + * the whole frame, therefore we don't submit the bitstream here */ + gst_v4l2_codec_h265_dec_fill_slice_params (self, slice, picture); + gst_v4l2_codec_h265_dec_fill_references (self, ref_pic_list0, + ref_pic_list1); + } + + /* if it is the first slice segment provide the short and long term + * reference pictures set size */ + if (slice->header.first_slice_segment_in_pic_flag) { + struct v4l2_ctrl_hevc_decode_params *decode_params = &self->decode_params; + + decode_params->short_term_ref_pic_set_size = + slice->header.short_term_ref_pic_set_size; + decode_params->long_term_ref_pic_set_size = + slice->header.long_term_ref_pic_set_size; + } + + bitstream_data = self->bitstream_map.data + self->bitstream_map.size; + + if (needs_start_codes (self)) + sc_off = 3; + nal_size = sc_off + slice->nalu.size; + + if (self->bitstream_map.size + nal_size > self->bitstream_map.maxsize) { + GST_ELEMENT_ERROR (decoder, RESOURCE, NO_SPACE_LEFT, + ("Not enough space to send all slice of an H265 frame."), (NULL)); + return GST_FLOW_ERROR; + } + + if (needs_start_codes (self)) { + bitstream_data[0] = 0x00; + bitstream_data[1] = 0x00; + bitstream_data[2] = 0x01; + } + + memcpy (bitstream_data + sc_off, slice->nalu.data + slice->nalu.offset, + slice->nalu.size); + self->bitstream_map.size += nal_size; + + return GST_FLOW_OK; +} + +static GstFlowReturn +gst_v4l2_codec_h265_dec_end_picture (GstH265Decoder * decoder, + GstH265Picture * picture) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + + if (!gst_v4l2_codec_h265_dec_submit_bitstream (self, picture, 0)) + return GST_FLOW_ERROR; + + return GST_FLOW_OK; +} + +static guint +gst_v4l2_codec_h265_dec_get_preferred_output_delay (GstH265Decoder * decoder, + gboolean is_live) +{ + + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + guint delay; + + if (is_live) + delay = 0; + else + delay = 1; + + gst_v4l2_decoder_set_render_delay (self->decoder, delay); + return delay; +} + +static void +gst_v4l2_codec_h265_dec_set_flushing (GstV4l2CodecH265Dec * self, + gboolean flushing) +{ + if (self->sink_allocator) + gst_v4l2_codec_allocator_set_flushing (self->sink_allocator, flushing); + if (self->src_allocator) + gst_v4l2_codec_allocator_set_flushing (self->src_allocator, flushing); +} + +static gboolean +gst_v4l2_codec_h265_dec_flush (GstVideoDecoder * decoder) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + + GST_DEBUG_OBJECT (self, "Flushing decoder state."); + + gst_v4l2_decoder_flush (self->decoder); + gst_v4l2_codec_h265_dec_set_flushing (self, FALSE); + + return GST_VIDEO_DECODER_CLASS (parent_class)->flush (decoder); +} + +static gboolean +gst_v4l2_codec_h265_dec_sink_event (GstVideoDecoder * decoder, GstEvent * event) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (decoder); + + switch (GST_EVENT_TYPE (event)) { + case GST_EVENT_FLUSH_START: + GST_DEBUG_OBJECT (self, "flush start"); + gst_v4l2_codec_h265_dec_set_flushing (self, TRUE); + break; + default: + break; + } + + return GST_VIDEO_DECODER_CLASS (parent_class)->sink_event (decoder, event); +} + +static GstStateChangeReturn +gst_v4l2_codec_h265_dec_change_state (GstElement * element, + GstStateChange transition) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (element); + + if (transition == GST_STATE_CHANGE_PAUSED_TO_READY) + gst_v4l2_codec_h265_dec_set_flushing (self, TRUE); + + return GST_ELEMENT_CLASS (parent_class)->change_state (element, transition); +} + +static void +gst_v4l2_codec_h265_dec_set_property (GObject * object, guint prop_id, + const GValue * value, GParamSpec * pspec) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (object); + GObject *dec = G_OBJECT (self->decoder); + + switch (prop_id) { + default: + gst_v4l2_decoder_set_property (dec, prop_id - PROP_LAST, value, pspec); + break; + } +} + +static void +gst_v4l2_codec_h265_dec_get_property (GObject * object, guint prop_id, + GValue * value, GParamSpec * pspec) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (object); + GObject *dec = G_OBJECT (self->decoder); + + switch (prop_id) { + default: + gst_v4l2_decoder_get_property (dec, prop_id - PROP_LAST, value, pspec); + break; + } +} + +static void +gst_v4l2_codec_h265_dec_init (GstV4l2CodecH265Dec * self) +{ +} + +static void +gst_v4l2_codec_h265_dec_subinit (GstV4l2CodecH265Dec * self, + GstV4l2CodecH265DecClass * klass) +{ + self->decoder = gst_v4l2_decoder_new (klass->device); + gst_video_info_init (&self->vinfo); + self->slice_params = g_array_sized_new (FALSE, TRUE, + sizeof (struct v4l2_ctrl_hevc_slice_params), 4); + g_array_set_size (self->slice_params, 4); + self->entry_point_offsets = g_array_sized_new (FALSE, TRUE, + sizeof (guint32), 4); +} + +static void +gst_v4l2_codec_h265_dec_dispose (GObject * object) +{ + GstV4l2CodecH265Dec *self = GST_V4L2_CODEC_H265_DEC (object); + + g_clear_object (&self->decoder); + g_clear_pointer (&self->slice_params, g_array_unref); + + G_OBJECT_CLASS (parent_class)->dispose (object); +} + +static void +gst_v4l2_codec_h265_dec_class_init (GstV4l2CodecH265DecClass * klass) +{ +} + +static void +gst_v4l2_codec_h265_dec_subclass_init (GstV4l2CodecH265DecClass * klass, + GstV4l2CodecDevice * device) +{ + GObjectClass *gobject_class = G_OBJECT_CLASS (klass); + GstElementClass *element_class = GST_ELEMENT_CLASS (klass); + GstVideoDecoderClass *decoder_class = GST_VIDEO_DECODER_CLASS (klass); + GstH265DecoderClass *h265decoder_class = GST_H265_DECODER_CLASS (klass); + + gobject_class->set_property = gst_v4l2_codec_h265_dec_set_property; + gobject_class->get_property = gst_v4l2_codec_h265_dec_get_property; + gobject_class->dispose = gst_v4l2_codec_h265_dec_dispose; + + gst_element_class_set_static_metadata (element_class, + "V4L2 Stateless H.265 Video Decoder", + "Codec/Decoder/Video/Hardware", + "A V4L2 based H.265 video decoder", + "Nicolas Dufresne <nicolas.dufresne@collabora.com>"); + + gst_element_class_add_static_pad_template (element_class, &sink_template); + gst_element_class_add_static_pad_template (element_class, &src_template); + element_class->change_state = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_change_state); + + decoder_class->open = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_open); + decoder_class->close = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_close); + decoder_class->stop = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_stop); + decoder_class->negotiate = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_negotiate); + decoder_class->decide_allocation = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_decide_allocation); + decoder_class->flush = GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_flush); + decoder_class->sink_event = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_sink_event); + + h265decoder_class->new_sequence = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_new_sequence); + h265decoder_class->output_picture = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_output_picture); + h265decoder_class->start_picture = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_start_picture); + h265decoder_class->decode_slice = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_decode_slice); + h265decoder_class->end_picture = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_end_picture); + h265decoder_class->get_preferred_output_delay = + GST_DEBUG_FUNCPTR (gst_v4l2_codec_h265_dec_get_preferred_output_delay); + + klass->device = device; + gst_v4l2_decoder_install_properties (gobject_class, PROP_LAST, device); +} + +void +gst_v4l2_codec_h265_dec_register (GstPlugin * plugin, GstV4l2Decoder * decoder, + GstV4l2CodecDevice * device, guint rank) +{ + GstCaps *src_caps; + guint version; + + GST_DEBUG_CATEGORY_INIT (v4l2_h265dec_debug, "v4l2codecs-h265dec", 0, + "V4L2 stateless h265 decoder"); + + if (!gst_v4l2_decoder_set_sink_fmt (decoder, V4L2_PIX_FMT_HEVC_SLICE, + 320, 240, 8)) + return; + src_caps = gst_v4l2_decoder_enum_src_formats (decoder); + + if (gst_caps_is_empty (src_caps)) { + GST_WARNING ("Not registering H265 decoder since it produces no " + "supported format"); + goto done; + } + + version = gst_v4l2_decoder_get_version (decoder); + if (version < V4L2_MIN_KERNEL_VERSION) + GST_WARNING ("V4L2 API v%u.%u too old, at least v%u.%u required", + (version >> 16) & 0xff, (version >> 8) & 0xff, + V4L2_MIN_KERNEL_VER_MAJOR, V4L2_MIN_KERNEL_VER_MINOR); + + gst_v4l2_decoder_register (plugin, GST_TYPE_V4L2_CODEC_H265_DEC, + (GClassInitFunc) gst_v4l2_codec_h265_dec_subclass_init, + gst_mini_object_ref (GST_MINI_OBJECT (device)), + (GInstanceInitFunc) gst_v4l2_codec_h265_dec_subinit, + "v4l2sl%sh265dec", device, rank, NULL); + +done: + gst_caps_unref (src_caps); +} diff --git a/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.h b/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.h new file mode 100644 index 0000000000..2a73435cb3 --- /dev/null +++ b/subprojects/gst-plugins-bad/sys/v4l2codecs/gstv4l2codech265dec.h @@ -0,0 +1,55 @@ +/* GStreamer + * Copyright (C) 2020 Nicolas Dufresne <nicolas.dufresne@collabora.com> + * Copyright (C) 2020 Safran Passenger Innovations LLC + * + * This library is free software; you can redistribute it and/or + * modify it under the terms of the GNU Library General Public + * License as published by the Free Software Foundation; either + * version 2 of the License, or (at your option) any later version. + * + * This library is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Library General Public License for more details. + * + * You should have received a copy of the GNU Library General Public + * License along with this library; if not, write to the + * Free Software Foundation, Inc., 51 Franklin St, Fifth Floor, + * Boston, MA 02110-1301, USA. + */ + +#ifndef __GST_V4L2_CODEC_H265_DEC_H__ +#define __GST_V4L2_CODEC_H265_DEC_H__ + +#define GST_USE_UNSTABLE_API +#include <gst/codecs/gsth265decoder.h> + +#include "gstv4l2decoder.h" + +G_BEGIN_DECLS + +#define GST_TYPE_V4L2_CODEC_H265_DEC (gst_v4l2_codec_h265_dec_get_type()) +#define GST_V4L2_CODEC_H265_DEC(obj) (G_TYPE_CHECK_INSTANCE_CAST((obj),GST_TYPE_V4L2_CODEC_H265_DEC,GstV4l2CodecH265Dec)) +#define GST_V4L2_CODEC_H265_DEC_CLASS(klass) (G_TYPE_CHECK_CLASS_CAST((klass),GST_TYPE_V4L2_CODEC_H265_DEC,GstV4l2CodecH265DecClass)) +#define GST_V4L2_CODEC_H265_DEC_GET_CLASS(obj) (G_TYPE_INSTANCE_GET_CLASS ((obj), GST_TYPE_V4L2_CODEC_H265_DEC, GstV4l2CodecH265DecClass)) +#define GST_IS_V4L2_CODEC_H265_DEC(obj) (G_TYPE_CHECK_INSTANCE_TYPE((obj),GST_TYPE_V4L2_CODEC_H265_DEC)) +#define GST_IS_V4L2_CODEC_H265_DEC_CLASS(obj) (G_TYPE_CHECK_CLASS_TYPE((klass),GST_TYPE_V4L2_CODEC_H265_DEC)) + +typedef struct _GstV4l2CodecH265Dec GstV4l2CodecH265Dec; +typedef struct _GstV4l2CodecH265DecClass GstV4l2CodecH265DecClass; + +struct _GstV4l2CodecH265DecClass +{ + GstH265DecoderClass parent_class; + GstV4l2CodecDevice *device; +}; + +GType gst_v4l2_codec_h265_dec_get_type (void); +void gst_v4l2_codec_h265_dec_register (GstPlugin * plugin, + GstV4l2Decoder * decoder, + GstV4l2CodecDevice * device, + guint rank); + +G_END_DECLS + +#endif /* __GST_D3D11_H265_DEC_H__ */ diff --git a/subprojects/gst-plugins-bad/sys/v4l2codecs/meson.build b/subprojects/gst-plugins-bad/sys/v4l2codecs/meson.build index ea8c79ced6..6b44dcad7d 100644 --- a/subprojects/gst-plugins-bad/sys/v4l2codecs/meson.build +++ b/subprojects/gst-plugins-bad/sys/v4l2codecs/meson.build @@ -3,6 +3,7 @@ v4l2codecs_sources = [ 'gstv4l2codecallocator.c', 'gstv4l2codecdevice.c', 'gstv4l2codech264dec.c', + 'gstv4l2codech265dec.c', 'gstv4l2codecmpeg2dec.c', 'gstv4l2codecpool.c', 'gstv4l2codecvp8dec.c', diff --git a/subprojects/gst-plugins-bad/sys/v4l2codecs/plugin.c b/subprojects/gst-plugins-bad/sys/v4l2codecs/plugin.c index 7d0796efa5..ac83e1b9d4 100644 --- a/subprojects/gst-plugins-bad/sys/v4l2codecs/plugin.c +++ b/subprojects/gst-plugins-bad/sys/v4l2codecs/plugin.c @@ -24,6 +24,7 @@ #include "gstv4l2codecdevice.h" #include "gstv4l2codech264dec.h" +#include "gstv4l2codech265dec.h" #include "gstv4l2codecmpeg2dec.h" #include "gstv4l2codecvp8dec.h" #include "gstv4l2codecvp9dec.h" @@ -54,6 +55,13 @@ register_video_decoder (GstPlugin * plugin, GstV4l2CodecDevice * device) gst_v4l2_codec_h264_dec_register (plugin, decoder, device, GST_RANK_PRIMARY + 1); break; + case V4L2_PIX_FMT_HEVC_SLICE: + GST_INFO_OBJECT (decoder, "Registering %s as H265 Decoder", + device->name); + gst_v4l2_codec_h265_dec_register (plugin, decoder, device, + GST_RANK_PRIMARY + 1); + break; + case V4L2_PIX_FMT_VP8_FRAME: GST_INFO_OBJECT (decoder, "Registering %s as VP8 Decoder", device->name); |