/************************************************************************** * * Copyright 2010 Thomas Balling Sørensen. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #include "util/u_memory.h" #include "util/u_math.h" #include "util/u_debug.h" #include "util/u_video.h" #include "util/vl_vlc.h" #include "vl/vl_codec.h" #include "vdpau_private.h" #define AV1_KEY_FRAME 0 #define AV1_REFS_PER_FRAME 7 #define AV1_NUM_REF_FRAMES 8 #define AV1_PRIMARY_REF_NONE AV1_REFS_PER_FRAME #define AV1_SUPERRES_DENOM_MIN 9 #define AV1_SUPERRES_NUM 8 /** * Create a VdpDecoder. */ VdpStatus vlVdpDecoderCreate(VdpDevice device, VdpDecoderProfile profile, uint32_t width, uint32_t height, uint32_t max_references, VdpDecoder *decoder) { struct pipe_video_codec templat = {}; struct pipe_context *pipe; struct pipe_screen *screen; vlVdpDevice *dev; vlVdpDecoder *vldecoder; VdpStatus ret; bool supported; uint32_t maxwidth, maxheight; if (!decoder) return VDP_STATUS_INVALID_POINTER; *decoder = 0; if (!(width && height)) return VDP_STATUS_INVALID_VALUE; templat.profile = ProfileToPipe(profile); if (templat.profile == PIPE_VIDEO_PROFILE_UNKNOWN) return VDP_STATUS_INVALID_DECODER_PROFILE; dev = vlGetDataHTAB(device); if (!dev) return VDP_STATUS_INVALID_HANDLE; pipe = dev->context; screen = dev->vscreen->pscreen; mtx_lock(&dev->mutex); supported = vl_codec_supported(screen, templat.profile, false); if (!supported) { mtx_unlock(&dev->mutex); return VDP_STATUS_INVALID_DECODER_PROFILE; } maxwidth = screen->get_video_param ( screen, templat.profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM, PIPE_VIDEO_CAP_MAX_WIDTH ); maxheight = screen->get_video_param ( screen, templat.profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM, PIPE_VIDEO_CAP_MAX_HEIGHT ); if (width > maxwidth || height > maxheight) { mtx_unlock(&dev->mutex); return VDP_STATUS_INVALID_SIZE; } vldecoder = CALLOC(1,sizeof(vlVdpDecoder)); if (!vldecoder) { mtx_unlock(&dev->mutex); return VDP_STATUS_RESOURCES; } DeviceReference(&vldecoder->device, dev); templat.entrypoint = PIPE_VIDEO_ENTRYPOINT_BITSTREAM; templat.chroma_format = PIPE_VIDEO_CHROMA_FORMAT_420; templat.width = width; templat.height = height; templat.max_references = max_references; if (u_reduce_video_profile(templat.profile) == PIPE_VIDEO_FORMAT_MPEG4_AVC) templat.level = u_get_h264_level(templat.width, templat.height, &templat.max_references); vldecoder->decoder = pipe->create_video_codec(pipe, &templat); if (!vldecoder->decoder) { ret = VDP_STATUS_ERROR; goto error_decoder; } *decoder = vlAddDataHTAB(vldecoder); if (*decoder == 0) { ret = VDP_STATUS_ERROR; goto error_handle; } (void) mtx_init(&vldecoder->mutex, mtx_plain); mtx_unlock(&dev->mutex); return VDP_STATUS_OK; error_handle: vldecoder->decoder->destroy(vldecoder->decoder); error_decoder: mtx_unlock(&dev->mutex); DeviceReference(&vldecoder->device, NULL); FREE(vldecoder); return ret; } /** * Destroy a VdpDecoder. */ VdpStatus vlVdpDecoderDestroy(VdpDecoder decoder) { vlVdpDecoder *vldecoder; vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder); if (!vldecoder) return VDP_STATUS_INVALID_HANDLE; mtx_lock(&vldecoder->mutex); vldecoder->decoder->destroy(vldecoder->decoder); mtx_unlock(&vldecoder->mutex); mtx_destroy(&vldecoder->mutex); vlRemoveDataHTAB(decoder); DeviceReference(&vldecoder->device, NULL); FREE(vldecoder); return VDP_STATUS_OK; } /** * Retrieve the parameters used to create a VdpDecoder. */ VdpStatus vlVdpDecoderGetParameters(VdpDecoder decoder, VdpDecoderProfile *profile, uint32_t *width, uint32_t *height) { vlVdpDecoder *vldecoder; vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder); if (!vldecoder) return VDP_STATUS_INVALID_HANDLE; *profile = PipeToProfile(vldecoder->decoder->profile); *width = vldecoder->decoder->width; *height = vldecoder->decoder->height; return VDP_STATUS_OK; } static VdpStatus vlVdpGetReferenceFrame(VdpVideoSurface handle, struct pipe_video_buffer **ref_frame) { vlVdpSurface *surface; /* if surfaces equals VDP_STATUS_INVALID_HANDLE, they are not used */ if (handle == VDP_INVALID_HANDLE) { *ref_frame = NULL; return VDP_STATUS_OK; } surface = vlGetDataHTAB(handle); if (!surface) return VDP_STATUS_INVALID_HANDLE; *ref_frame = surface->video_buffer; if (!*ref_frame) return VDP_STATUS_INVALID_HANDLE; return VDP_STATUS_OK; } /** * Decode a mpeg 1/2 video. */ static VdpStatus vlVdpDecoderRenderMpeg12(struct pipe_mpeg12_picture_desc *picture, const VdpPictureInfoMPEG1Or2 *picture_info) { VdpStatus r; VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding MPEG12\n"); r = vlVdpGetReferenceFrame(picture_info->forward_reference, &picture->ref[0]); if (r != VDP_STATUS_OK) return r; r = vlVdpGetReferenceFrame(picture_info->backward_reference, &picture->ref[1]); if (r != VDP_STATUS_OK) return r; picture->picture_coding_type = picture_info->picture_coding_type; picture->picture_structure = picture_info->picture_structure; picture->frame_pred_frame_dct = picture_info->frame_pred_frame_dct; picture->q_scale_type = picture_info->q_scale_type; picture->alternate_scan = picture_info->alternate_scan; picture->intra_vlc_format = picture_info->intra_vlc_format; picture->concealment_motion_vectors = picture_info->concealment_motion_vectors; picture->intra_dc_precision = picture_info->intra_dc_precision; picture->f_code[0][0] = picture_info->f_code[0][0] - 1; picture->f_code[0][1] = picture_info->f_code[0][1] - 1; picture->f_code[1][0] = picture_info->f_code[1][0] - 1; picture->f_code[1][1] = picture_info->f_code[1][1] - 1; picture->num_slices = picture_info->slice_count; picture->top_field_first = picture_info->top_field_first; picture->full_pel_forward_vector = picture_info->full_pel_forward_vector; picture->full_pel_backward_vector = picture_info->full_pel_backward_vector; picture->intra_matrix = picture_info->intra_quantizer_matrix; picture->non_intra_matrix = picture_info->non_intra_quantizer_matrix; return VDP_STATUS_OK; } /** * Decode a mpeg 4 video. */ static VdpStatus vlVdpDecoderRenderMpeg4(struct pipe_mpeg4_picture_desc *picture, const VdpPictureInfoMPEG4Part2 *picture_info) { VdpStatus r; unsigned i; VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding MPEG4\n"); r = vlVdpGetReferenceFrame(picture_info->forward_reference, &picture->ref[0]); if (r != VDP_STATUS_OK) return r; r = vlVdpGetReferenceFrame(picture_info->backward_reference, &picture->ref[1]); if (r != VDP_STATUS_OK) return r; for (i = 0; i < 2; ++i) { picture->trd[i] = picture_info->trd[i]; picture->trb[i] = picture_info->trb[i]; } picture->vop_time_increment_resolution = picture_info->vop_time_increment_resolution; picture->vop_coding_type = picture_info->vop_coding_type; picture->vop_fcode_forward = picture_info->vop_fcode_forward; picture->vop_fcode_backward = picture_info->vop_fcode_backward; picture->resync_marker_disable = picture_info->resync_marker_disable; picture->interlaced = picture_info->interlaced; picture->quant_type = picture_info->quant_type; picture->quarter_sample = picture_info->quarter_sample; picture->short_video_header = picture_info->short_video_header; picture->rounding_control = picture_info->rounding_control; picture->alternate_vertical_scan_flag = picture_info->alternate_vertical_scan_flag; picture->top_field_first = picture_info->top_field_first; picture->intra_matrix = picture_info->intra_quantizer_matrix; picture->non_intra_matrix = picture_info->non_intra_quantizer_matrix; return VDP_STATUS_OK; } static VdpStatus vlVdpDecoderRenderVC1(struct pipe_vc1_picture_desc *picture, const VdpPictureInfoVC1 *picture_info) { VdpStatus r; VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding VC-1\n"); r = vlVdpGetReferenceFrame(picture_info->forward_reference, &picture->ref[0]); if (r != VDP_STATUS_OK) return r; r = vlVdpGetReferenceFrame(picture_info->backward_reference, &picture->ref[1]); if (r != VDP_STATUS_OK) return r; picture->slice_count = picture_info->slice_count; picture->picture_type = picture_info->picture_type; picture->frame_coding_mode = picture_info->frame_coding_mode; picture->postprocflag = picture_info->postprocflag; picture->pulldown = picture_info->pulldown; picture->interlace = picture_info->interlace; picture->tfcntrflag = picture_info->tfcntrflag; picture->finterpflag = picture_info->finterpflag; picture->psf = picture_info->psf; picture->dquant = picture_info->dquant; picture->panscan_flag = picture_info->panscan_flag; picture->refdist_flag = picture_info->refdist_flag; picture->quantizer = picture_info->quantizer; picture->extended_mv = picture_info->extended_mv; picture->extended_dmv = picture_info->extended_dmv; picture->overlap = picture_info->overlap; picture->vstransform = picture_info->vstransform; picture->loopfilter = picture_info->loopfilter; picture->fastuvmc = picture_info->fastuvmc; picture->range_mapy_flag = picture_info->range_mapy_flag; picture->range_mapy = picture_info->range_mapy; picture->range_mapuv_flag = picture_info->range_mapuv_flag; picture->range_mapuv = picture_info->range_mapuv; picture->multires = picture_info->multires; picture->syncmarker = picture_info->syncmarker; picture->rangered = picture_info->rangered; picture->maxbframes = picture_info->maxbframes; picture->deblockEnable = picture_info->deblockEnable; picture->pquant = picture_info->pquant; return VDP_STATUS_OK; } static VdpStatus vlVdpDecoderRenderH264(struct pipe_h264_picture_desc *picture, const VdpPictureInfoH264 *picture_info, unsigned level_idc) { unsigned i; VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Decoding H264\n"); picture->pps->sps->mb_adaptive_frame_field_flag = picture_info->mb_adaptive_frame_field_flag; picture->pps->sps->frame_mbs_only_flag = picture_info->frame_mbs_only_flag; picture->pps->sps->log2_max_frame_num_minus4 = picture_info->log2_max_frame_num_minus4; picture->pps->sps->pic_order_cnt_type = picture_info->pic_order_cnt_type; picture->pps->sps->log2_max_pic_order_cnt_lsb_minus4 = picture_info->log2_max_pic_order_cnt_lsb_minus4; picture->pps->sps->delta_pic_order_always_zero_flag = picture_info->delta_pic_order_always_zero_flag; picture->pps->sps->direct_8x8_inference_flag = picture_info->direct_8x8_inference_flag; picture->pps->sps->level_idc = level_idc; picture->pps->sps->MinLumaBiPredSize8x8 = (level_idc >= 31); /* See section A.3.3.2 of H264 spec */; picture->pps->transform_8x8_mode_flag = picture_info->transform_8x8_mode_flag; picture->pps->chroma_qp_index_offset = picture_info->chroma_qp_index_offset; picture->pps->second_chroma_qp_index_offset = picture_info->second_chroma_qp_index_offset; picture->pps->pic_init_qp_minus26 = picture_info->pic_init_qp_minus26; /*picture->pps-> pic_init_qs_minus26 not passed in VdpPictureInfoH264*/ picture->pps->entropy_coding_mode_flag = picture_info->entropy_coding_mode_flag; picture->pps->deblocking_filter_control_present_flag = picture_info->deblocking_filter_control_present_flag; picture->pps->redundant_pic_cnt_present_flag = picture_info->redundant_pic_cnt_present_flag; picture->pps->constrained_intra_pred_flag = picture_info->constrained_intra_pred_flag; picture->pps->weighted_pred_flag = picture_info->weighted_pred_flag; picture->pps->weighted_bipred_idc = picture_info->weighted_bipred_idc; picture->pps->bottom_field_pic_order_in_frame_present_flag = picture_info->pic_order_present_flag; memcpy(picture->pps->ScalingList4x4, picture_info->scaling_lists_4x4, 6*16); memcpy(picture->pps->ScalingList8x8, picture_info->scaling_lists_8x8, 2*64); picture->slice_count = picture_info->slice_count; picture->field_order_cnt[0] = picture_info->field_order_cnt[0]; picture->field_order_cnt[1] = picture_info->field_order_cnt[1]; picture->is_reference = picture_info->is_reference; picture->frame_num = picture_info->frame_num; picture->field_pic_flag = picture_info->field_pic_flag; picture->bottom_field_flag = picture_info->bottom_field_flag; picture->num_ref_frames = picture_info->num_ref_frames; picture->num_ref_idx_l0_active_minus1 = picture_info->num_ref_idx_l0_active_minus1; picture->num_ref_idx_l1_active_minus1 = picture_info->num_ref_idx_l1_active_minus1; for (i = 0; i < 16; ++i) { VdpStatus ret = vlVdpGetReferenceFrame ( picture_info->referenceFrames[i].surface, &picture->ref[i] ); if (ret != VDP_STATUS_OK) return ret; picture->is_long_term[i] = picture_info->referenceFrames[i].is_long_term; picture->top_is_reference[i] = picture_info->referenceFrames[i].top_is_reference; picture->bottom_is_reference[i] = picture_info->referenceFrames[i].bottom_is_reference; picture->field_order_cnt_list[i][0] = picture_info->referenceFrames[i].field_order_cnt[0]; picture->field_order_cnt_list[i][1] = picture_info->referenceFrames[i].field_order_cnt[1]; picture->frame_num_list[i] = picture_info->referenceFrames[i].frame_idx; } return VDP_STATUS_OK; } static VdpStatus vlVdpDecoderRenderH265(struct pipe_h265_picture_desc *picture, const VdpPictureInfoHEVC *picture_info) { unsigned i; picture->pps->sps->chroma_format_idc = picture_info->chroma_format_idc; picture->pps->sps->separate_colour_plane_flag = picture_info->separate_colour_plane_flag; picture->pps->sps->pic_width_in_luma_samples = picture_info->pic_width_in_luma_samples; picture->pps->sps->pic_height_in_luma_samples = picture_info->pic_height_in_luma_samples; picture->pps->sps->bit_depth_luma_minus8 = picture_info->bit_depth_luma_minus8; picture->pps->sps->bit_depth_chroma_minus8 = picture_info->bit_depth_chroma_minus8; picture->pps->sps->log2_max_pic_order_cnt_lsb_minus4 = picture_info->log2_max_pic_order_cnt_lsb_minus4; picture->pps->sps->sps_max_dec_pic_buffering_minus1 = picture_info->sps_max_dec_pic_buffering_minus1; picture->pps->sps->log2_min_luma_coding_block_size_minus3 = picture_info->log2_min_luma_coding_block_size_minus3; picture->pps->sps->log2_diff_max_min_luma_coding_block_size = picture_info->log2_diff_max_min_luma_coding_block_size; picture->pps->sps->log2_min_transform_block_size_minus2 = picture_info->log2_min_transform_block_size_minus2; picture->pps->sps->log2_diff_max_min_transform_block_size = picture_info->log2_diff_max_min_transform_block_size; picture->pps->sps->max_transform_hierarchy_depth_inter = picture_info->max_transform_hierarchy_depth_inter; picture->pps->sps->max_transform_hierarchy_depth_intra = picture_info->max_transform_hierarchy_depth_intra; picture->pps->sps->scaling_list_enabled_flag = picture_info->scaling_list_enabled_flag; memcpy(picture->pps->sps->ScalingList4x4, picture_info->ScalingList4x4, 6*16); memcpy(picture->pps->sps->ScalingList8x8, picture_info->ScalingList8x8, 6*64); memcpy(picture->pps->sps->ScalingList16x16, picture_info->ScalingList16x16, 6*64); memcpy(picture->pps->sps->ScalingList32x32, picture_info->ScalingList32x32, 2*64); memcpy(picture->pps->sps->ScalingListDCCoeff16x16, picture_info->ScalingListDCCoeff16x16, 6); memcpy(picture->pps->sps->ScalingListDCCoeff32x32, picture_info->ScalingListDCCoeff32x32, 2); picture->pps->sps->amp_enabled_flag = picture_info->amp_enabled_flag; picture->pps->sps->sample_adaptive_offset_enabled_flag = picture_info->sample_adaptive_offset_enabled_flag; picture->pps->sps->pcm_enabled_flag = picture_info->pcm_enabled_flag; picture->pps->sps->pcm_sample_bit_depth_luma_minus1 = picture_info->pcm_sample_bit_depth_luma_minus1; picture->pps->sps->pcm_sample_bit_depth_chroma_minus1 = picture_info->pcm_sample_bit_depth_chroma_minus1; picture->pps->sps->log2_min_pcm_luma_coding_block_size_minus3 = picture_info->log2_min_pcm_luma_coding_block_size_minus3; picture->pps->sps->log2_diff_max_min_pcm_luma_coding_block_size = picture_info->log2_diff_max_min_pcm_luma_coding_block_size; picture->pps->sps->pcm_loop_filter_disabled_flag = picture_info->pcm_loop_filter_disabled_flag; picture->pps->sps->num_short_term_ref_pic_sets = picture_info->num_short_term_ref_pic_sets; picture->pps->sps->long_term_ref_pics_present_flag = picture_info->long_term_ref_pics_present_flag; picture->pps->sps->num_long_term_ref_pics_sps = picture_info->num_long_term_ref_pics_sps; picture->pps->sps->sps_temporal_mvp_enabled_flag = picture_info->sps_temporal_mvp_enabled_flag; picture->pps->sps->strong_intra_smoothing_enabled_flag = picture_info->strong_intra_smoothing_enabled_flag; picture->pps->dependent_slice_segments_enabled_flag = picture_info->dependent_slice_segments_enabled_flag; picture->pps->output_flag_present_flag = picture_info->output_flag_present_flag; picture->pps->num_extra_slice_header_bits = picture_info->num_extra_slice_header_bits; picture->pps->sign_data_hiding_enabled_flag = picture_info->sign_data_hiding_enabled_flag; picture->pps->cabac_init_present_flag = picture_info->cabac_init_present_flag; picture->pps->num_ref_idx_l0_default_active_minus1 = picture_info->num_ref_idx_l0_default_active_minus1; picture->pps->num_ref_idx_l1_default_active_minus1 = picture_info->num_ref_idx_l1_default_active_minus1; picture->pps->init_qp_minus26 = picture_info->init_qp_minus26; picture->pps->constrained_intra_pred_flag = picture_info->constrained_intra_pred_flag; picture->pps->transform_skip_enabled_flag = picture_info->transform_skip_enabled_flag; picture->pps->cu_qp_delta_enabled_flag = picture_info->cu_qp_delta_enabled_flag; picture->pps->diff_cu_qp_delta_depth = picture_info->diff_cu_qp_delta_depth; picture->pps->pps_cb_qp_offset = picture_info->pps_cb_qp_offset; picture->pps->pps_cr_qp_offset = picture_info->pps_cr_qp_offset; picture->pps->pps_slice_chroma_qp_offsets_present_flag = picture_info->pps_slice_chroma_qp_offsets_present_flag; picture->pps->weighted_pred_flag = picture_info->weighted_pred_flag; picture->pps->weighted_bipred_flag = picture_info->weighted_bipred_flag; picture->pps->transquant_bypass_enabled_flag = picture_info->transquant_bypass_enabled_flag; picture->pps->tiles_enabled_flag = picture_info->tiles_enabled_flag; picture->pps->entropy_coding_sync_enabled_flag = picture_info->entropy_coding_sync_enabled_flag; picture->pps->num_tile_columns_minus1 = picture_info->num_tile_columns_minus1; picture->pps->num_tile_rows_minus1 = picture_info->num_tile_rows_minus1; picture->pps->uniform_spacing_flag = picture_info->uniform_spacing_flag; memcpy(picture->pps->column_width_minus1, picture_info->column_width_minus1, 20 * 2); memcpy(picture->pps->row_height_minus1, picture_info->row_height_minus1, 22 * 2); picture->pps->loop_filter_across_tiles_enabled_flag = picture_info->loop_filter_across_tiles_enabled_flag; picture->pps->pps_loop_filter_across_slices_enabled_flag = picture_info->pps_loop_filter_across_slices_enabled_flag; picture->pps->deblocking_filter_control_present_flag = picture_info->deblocking_filter_control_present_flag; picture->pps->deblocking_filter_override_enabled_flag = picture_info->deblocking_filter_override_enabled_flag; picture->pps->pps_deblocking_filter_disabled_flag = picture_info->pps_deblocking_filter_disabled_flag; picture->pps->pps_beta_offset_div2 = picture_info->pps_beta_offset_div2; picture->pps->pps_tc_offset_div2 = picture_info->pps_tc_offset_div2; picture->pps->lists_modification_present_flag = picture_info->lists_modification_present_flag; picture->pps->log2_parallel_merge_level_minus2 = picture_info->log2_parallel_merge_level_minus2; picture->pps->slice_segment_header_extension_present_flag = picture_info->slice_segment_header_extension_present_flag; picture->IDRPicFlag = picture_info->IDRPicFlag; picture->RAPPicFlag = picture_info->RAPPicFlag; picture->IntraPicFlag = picture_info->RAPPicFlag; picture->CurrRpsIdx = picture_info->CurrRpsIdx; picture->NumPocTotalCurr = picture_info->NumPocTotalCurr; picture->NumDeltaPocsOfRefRpsIdx = picture_info->NumDeltaPocsOfRefRpsIdx; picture->NumShortTermPictureSliceHeaderBits = picture_info->NumShortTermPictureSliceHeaderBits; picture->NumLongTermPictureSliceHeaderBits = picture_info->NumLongTermPictureSliceHeaderBits; picture->CurrPicOrderCntVal = picture_info->CurrPicOrderCntVal; for (i = 0; i < 16; ++i) { VdpStatus ret = vlVdpGetReferenceFrame ( picture_info->RefPics[i], &picture->ref[i] ); if (ret != VDP_STATUS_OK) return ret; picture->PicOrderCntVal[i] = picture_info->PicOrderCntVal[i]; picture->IsLongTerm[i] = picture_info->IsLongTerm[i]; } picture->NumPocStCurrBefore = picture_info->NumPocStCurrBefore; picture->NumPocStCurrAfter = picture_info->NumPocStCurrAfter; picture->NumPocLtCurr = picture_info->NumPocLtCurr; memcpy(picture->RefPicSetStCurrBefore, picture_info->RefPicSetStCurrBefore, 8); memcpy(picture->RefPicSetStCurrAfter, picture_info->RefPicSetStCurrAfter, 8); memcpy(picture->RefPicSetLtCurr, picture_info->RefPicSetLtCurr, 8); picture->UseRefPicList = false; picture->UseStRpsBits = false; return VDP_STATUS_OK; } static void copyArrayInt8FromShort(int8_t *dest, const short *src, unsigned count) { unsigned i; for (i = 0; i < count; ++i) { *dest = *src; ++dest; ++src; } } static void copyAV1ScalingPoints(uint8_t *value, uint8_t *scaling, const unsigned char point[][2], unsigned count) { unsigned i; for (i = 0; i < count; ++i) { *value = (*point)[0]; ++value; *scaling = (*point)[1]; ++scaling; ++point; } } static uint8_t indexOfAV1RefFrame(uint32_t frame, const unsigned int *ref_frame_map) { uint8_t i; for (i = 0; i < AV1_NUM_REF_FRAMES; ++i) { if (frame == *ref_frame_map) { break; } ++ref_frame_map; } return i; } static void copyAV1TileInfo(struct pipe_av1_picture_desc *picture, const VdpPictureInfoAV1 *picture_info) { uint32_t sbCols, sbRows; uint32_t startSb, i; int32_t width_sb, height_sb; uint32_t MiCols = ((picture_info->width + 7) >> 3) << 1; uint32_t MiRows = ((picture_info->height + 7) >> 3) << 1; if (picture_info->use_superres) { const uint32_t superres_scale_denominator = picture_info->coded_denom + AV1_SUPERRES_DENOM_MIN; const uint32_t width = ((picture_info->width * 8) + (superres_scale_denominator / 2)) / superres_scale_denominator; MiCols = (((width - 1) + 8) >> 3) << 1; } sbCols = picture_info->use_128x128_superblock ? ((MiCols + 31) >> 5) : ((MiCols + 15) >> 4); sbRows = picture_info->use_128x128_superblock ? ((MiRows + 31) >> 5) : ((MiRows + 15) >> 4); width_sb = sbCols; height_sb = sbRows; startSb = 0; for (i = 0; startSb < sbCols; ++i) { const uint32_t tile_width = picture_info->tile_widths[i]; picture->picture_parameter.width_in_sbs[i] = tile_width; picture->picture_parameter.tile_col_start_sb[i] = startSb; startSb += tile_width; width_sb -= tile_width; } picture->picture_parameter.tile_col_start_sb[i] = startSb + width_sb; startSb = 0; for (i = 0; startSb < sbRows; ++i) { const uint32_t tile_height = picture_info->tile_heights[i]; picture->picture_parameter.height_in_sbs[i] = tile_height; picture->picture_parameter.tile_row_start_sb[i] = startSb; startSb += tile_height; height_sb -= tile_height; } picture->picture_parameter.tile_row_start_sb[i] = startSb + height_sb; } static VdpStatus vlVdpDecoderRenderAV1(struct pipe_av1_picture_desc *picture, VdpVideoSurface target, const VdpPictureInfoAV1 *picture_info) { unsigned i, j; picture->film_grain_target = NULL; picture->picture_parameter.profile = picture_info->profile; picture->picture_parameter.order_hint_bits_minus_1 = picture_info->order_hint_bits_minus1; picture->picture_parameter.bit_depth_idx = picture_info->bit_depth_minus8 >> 1; picture->picture_parameter.seq_info_fields.use_128x128_superblock = picture_info->use_128x128_superblock; picture->picture_parameter.seq_info_fields.enable_filter_intra = picture_info->enable_filter_intra; picture->picture_parameter.seq_info_fields.enable_intra_edge_filter = picture_info->enable_intra_edge_filter; picture->picture_parameter.seq_info_fields.enable_interintra_compound = picture_info->enable_interintra_compound; picture->picture_parameter.seq_info_fields.enable_masked_compound = picture_info->enable_masked_compound; picture->picture_parameter.seq_info_fields.enable_dual_filter = picture_info->enable_dual_filter; picture->picture_parameter.seq_info_fields.enable_order_hint = picture_info->enable_order_hint; picture->picture_parameter.seq_info_fields.enable_jnt_comp = picture_info->enable_jnt_comp; picture->picture_parameter.seq_info_fields.enable_cdef = picture_info->enable_cdef; picture->picture_parameter.seq_info_fields.mono_chrome = picture_info->mono_chrome; picture->picture_parameter.seq_info_fields.ref_frame_mvs = picture_info->enable_order_hint; picture->picture_parameter.seq_info_fields.film_grain_params_present = picture_info->enable_fgs; picture->picture_parameter.current_frame_id = target; picture->picture_parameter.frame_width = picture_info->width; picture->picture_parameter.frame_height = picture_info->height; picture->picture_parameter.max_width = picture_info->width; picture->picture_parameter.max_height = picture_info->height; for (i = 0; i < AV1_NUM_REF_FRAMES; ++i) { if (picture_info->frame_type == AV1_KEY_FRAME && picture_info->show_frame) { picture->ref[i] = NULL; } else { VdpStatus ret = vlVdpGetReferenceFrame(picture_info->ref_frame_map[i], &picture->ref[i]); if (ret != VDP_STATUS_OK) { return ret; } } } for (i = 0; i < AV1_REFS_PER_FRAME; ++i) { const uint8_t idx = indexOfAV1RefFrame(picture_info->ref_frame[i].index, picture_info->ref_frame_map); if (idx == AV1_NUM_REF_FRAMES) { return VDP_STATUS_ERROR; } picture->picture_parameter.ref_frame_idx[i] = idx; } if (picture_info->primary_ref_frame == VDP_INVALID_HANDLE) { picture->picture_parameter.primary_ref_frame = AV1_PRIMARY_REF_NONE; } else { const uint8_t *ref_index = picture->picture_parameter.ref_frame_idx; const uint8_t idx = indexOfAV1RefFrame(picture_info->primary_ref_frame, picture_info->ref_frame_map); if (idx == AV1_NUM_REF_FRAMES) { return VDP_STATUS_ERROR; } for (i = 0; i < AV1_REFS_PER_FRAME; ++i) { if (idx == *ref_index) { break; } ++ref_index; } picture->picture_parameter.primary_ref_frame = i; } picture->picture_parameter.refresh_frame_flags = 0x01; picture->picture_parameter.order_hint = picture_info->frame_offset; // Segment Info picture->picture_parameter.seg_info.segment_info_fields.enabled = picture_info->segmentation_enabled; picture->picture_parameter.seg_info.segment_info_fields.update_map = picture_info->segmentation_update_map; picture->picture_parameter.seg_info.segment_info_fields.update_data = picture_info->segmentation_update_data; picture->picture_parameter.seg_info.segment_info_fields.temporal_update = picture_info->segmentation_temporal_update; memcpy(picture->picture_parameter.seg_info.feature_data, picture_info->segmentation_feature_data, sizeof(picture->picture_parameter.seg_info.feature_data)); memcpy(picture->picture_parameter.seg_info.feature_mask, picture_info->segmentation_feature_mask, sizeof(picture->picture_parameter.seg_info.feature_mask)); // Film Grain Info if (picture_info->enable_fgs) { picture->picture_parameter.film_grain_info.film_grain_info_fields.apply_grain = picture_info->apply_grain; picture->picture_parameter.film_grain_info.film_grain_info_fields.chroma_scaling_from_luma = picture_info->chroma_scaling_from_luma; picture->picture_parameter.film_grain_info.film_grain_info_fields.grain_scaling_minus_8 = picture_info->scaling_shift_minus8; picture->picture_parameter.film_grain_info.film_grain_info_fields.ar_coeff_lag = picture_info->ar_coeff_lag; picture->picture_parameter.film_grain_info.film_grain_info_fields.ar_coeff_shift_minus_6 = picture_info->ar_coeff_shift_minus6; picture->picture_parameter.film_grain_info.film_grain_info_fields.grain_scale_shift = picture_info->grain_scale_shift; picture->picture_parameter.film_grain_info.film_grain_info_fields.overlap_flag = picture_info->overlap_flag; picture->picture_parameter.film_grain_info.film_grain_info_fields.clip_to_restricted_range = picture_info->clip_to_restricted_range; picture->picture_parameter.film_grain_info.grain_seed = picture_info->random_seed; picture->picture_parameter.film_grain_info.num_y_points = picture_info->num_y_points; picture->picture_parameter.film_grain_info.num_cb_points = picture_info->num_cb_points; picture->picture_parameter.film_grain_info.num_cr_points = picture_info->num_cr_points; picture->picture_parameter.film_grain_info.cb_mult = picture_info->cb_mult; picture->picture_parameter.film_grain_info.cb_luma_mult = picture_info->cb_luma_mult; picture->picture_parameter.film_grain_info.cb_offset = picture_info->cb_offset; picture->picture_parameter.film_grain_info.cr_mult = picture_info->cr_mult; picture->picture_parameter.film_grain_info.cr_luma_mult = picture_info->cr_luma_mult; picture->picture_parameter.film_grain_info.cr_offset = picture_info->cr_offset; copyAV1ScalingPoints( picture->picture_parameter.film_grain_info.point_y_value, picture->picture_parameter.film_grain_info.point_y_scaling, picture_info->scaling_points_y, ARRAY_SIZE(picture->picture_parameter.film_grain_info.point_y_value)); copyAV1ScalingPoints( picture->picture_parameter.film_grain_info.point_cb_value, picture->picture_parameter.film_grain_info.point_cb_scaling, picture_info->scaling_points_cb, ARRAY_SIZE(picture->picture_parameter.film_grain_info.point_cb_value)); copyAV1ScalingPoints( picture->picture_parameter.film_grain_info.point_cr_value, picture->picture_parameter.film_grain_info.point_cr_scaling, picture_info->scaling_points_cr, ARRAY_SIZE(picture->picture_parameter.film_grain_info.point_cr_value)); copyArrayInt8FromShort( picture->picture_parameter.film_grain_info.ar_coeffs_y, picture_info->ar_coeffs_y, ARRAY_SIZE(picture->picture_parameter.film_grain_info.ar_coeffs_y)); copyArrayInt8FromShort( picture->picture_parameter.film_grain_info.ar_coeffs_cb, picture_info->ar_coeffs_cb, ARRAY_SIZE(picture->picture_parameter.film_grain_info.ar_coeffs_cb)); copyArrayInt8FromShort( picture->picture_parameter.film_grain_info.ar_coeffs_cr, picture_info->ar_coeffs_cr, ARRAY_SIZE(picture->picture_parameter.film_grain_info.ar_coeffs_cr)); } // Picture Info picture->picture_parameter.pic_info_fields.frame_type = picture_info->frame_type; picture->picture_parameter.pic_info_fields.show_frame = picture_info->show_frame; picture->picture_parameter.pic_info_fields.showable_frame = 1; picture->picture_parameter.pic_info_fields.error_resilient_mode = 1; picture->picture_parameter.pic_info_fields.disable_cdf_update = picture_info->disable_cdf_update; picture->picture_parameter.pic_info_fields.allow_screen_content_tools = picture_info->allow_screen_content_tools; picture->picture_parameter.pic_info_fields.force_integer_mv = picture_info->force_integer_mv; picture->picture_parameter.pic_info_fields.allow_intrabc = picture_info->allow_intrabc; picture->picture_parameter.pic_info_fields.use_superres = picture_info->use_superres; picture->picture_parameter.pic_info_fields.allow_high_precision_mv = picture_info->allow_high_precision_mv; picture->picture_parameter.pic_info_fields.is_motion_mode_switchable = picture_info->switchable_motion_mode; picture->picture_parameter.pic_info_fields.use_ref_frame_mvs = picture_info->use_ref_frame_mvs; picture->picture_parameter.pic_info_fields.disable_frame_end_update_cdf = picture_info->disable_frame_end_update_cdf; picture->picture_parameter.pic_info_fields.uniform_tile_spacing_flag = 0; picture->picture_parameter.pic_info_fields.allow_warped_motion = picture_info->allow_warped_motion; picture->picture_parameter.pic_info_fields.large_scale_tile = 0; picture->picture_parameter.superres_scale_denominator = picture_info->use_superres ? picture_info->coded_denom + AV1_SUPERRES_DENOM_MIN : AV1_SUPERRES_NUM; // Loop Filter picture->picture_parameter.interp_filter = picture_info->interp_filter; memcpy(picture->picture_parameter.filter_level, picture_info->loop_filter_level, sizeof(picture->picture_parameter.filter_level)); picture->picture_parameter.filter_level_u = picture_info->loop_filter_level_u; picture->picture_parameter.filter_level_v = picture_info->loop_filter_level_v; picture->picture_parameter.loop_filter_info_fields.sharpness_level = picture_info->loop_filter_sharpness; picture->picture_parameter.loop_filter_info_fields.mode_ref_delta_enabled = picture_info->loop_filter_delta_enabled; picture->picture_parameter.loop_filter_info_fields.mode_ref_delta_update = picture_info->loop_filter_delta_update; memcpy(picture->picture_parameter.ref_deltas, picture_info->loop_filter_ref_deltas, sizeof(picture->picture_parameter.ref_deltas)); memcpy(picture->picture_parameter.mode_deltas, picture_info->loop_filter_mode_deltas, sizeof(picture->picture_parameter.mode_deltas)); // Tile Info picture->picture_parameter.tile_cols = picture_info->num_tile_cols; picture->picture_parameter.tile_rows = picture_info->num_tile_rows; picture->picture_parameter.context_update_tile_id = picture_info->context_update_tile_id; copyAV1TileInfo(picture, picture_info); // Quantization Parameters picture->picture_parameter.base_qindex = picture_info->base_qindex; picture->picture_parameter.y_dc_delta_q = picture_info->qp_y_dc_delta_q; picture->picture_parameter.u_dc_delta_q = picture_info->qp_u_dc_delta_q; picture->picture_parameter.u_ac_delta_q = picture_info->qp_u_ac_delta_q; picture->picture_parameter.v_dc_delta_q = picture_info->qp_v_dc_delta_q; picture->picture_parameter.v_ac_delta_q = picture_info->qp_v_ac_delta_q; // QMatrix picture->picture_parameter.qmatrix_fields.using_qmatrix = picture_info->using_qmatrix; if (picture_info->using_qmatrix) { picture->picture_parameter.qmatrix_fields.qm_y = picture_info->qm_y; picture->picture_parameter.qmatrix_fields.qm_u = picture_info->qm_u; picture->picture_parameter.qmatrix_fields.qm_v = picture_info->qm_v; } else { picture->picture_parameter.qmatrix_fields.qm_y = 0x0f; picture->picture_parameter.qmatrix_fields.qm_u = 0x0f; picture->picture_parameter.qmatrix_fields.qm_v = 0x0f; } // Mode Control Fields picture->picture_parameter.mode_control_fields.delta_q_present_flag = picture_info->delta_q_present; picture->picture_parameter.mode_control_fields.log2_delta_q_res = picture_info->delta_q_res; picture->picture_parameter.mode_control_fields.delta_lf_present_flag = picture_info->delta_lf_present; picture->picture_parameter.mode_control_fields.log2_delta_lf_res = picture_info->delta_lf_res; picture->picture_parameter.mode_control_fields.delta_lf_multi = picture_info->delta_lf_multi; picture->picture_parameter.mode_control_fields.tx_mode = picture_info->tx_mode; picture->picture_parameter.mode_control_fields.reference_select = picture_info->reference_mode; picture->picture_parameter.mode_control_fields.reduced_tx_set_used = picture_info->reduced_tx_set; picture->picture_parameter.mode_control_fields.skip_mode_present = picture_info->skip_mode; // CDEF picture->picture_parameter.cdef_damping_minus_3 = picture_info->cdef_damping_minus_3; picture->picture_parameter.cdef_bits = picture_info->cdef_bits; for (i = 0; i < ARRAY_SIZE(picture->picture_parameter.cdef_y_strengths); ++i) { picture->picture_parameter.cdef_y_strengths[i] = ((picture_info->cdef_y_strength[i] & 0xf) << 2) + (picture_info->cdef_y_strength[i] >> 4); picture->picture_parameter.cdef_uv_strengths[i] = ((picture_info->cdef_uv_strength[i] & 0xf) << 2) + (picture_info->cdef_uv_strength[i] >> 4); } // Loop Restoration picture->picture_parameter.loop_restoration_fields.yframe_restoration_type = picture_info->lr_type[0]; picture->picture_parameter.loop_restoration_fields.cbframe_restoration_type = picture_info->lr_type[1]; picture->picture_parameter.loop_restoration_fields.crframe_restoration_type = picture_info->lr_type[2]; picture->picture_parameter.loop_restoration_fields.lr_unit_shift = picture_info->lr_unit_size[0] - 1; picture->picture_parameter.loop_restoration_fields.lr_uv_shift = picture_info->lr_unit_size[0] - picture_info->lr_unit_size[1]; if (picture_info->lr_type[0] || picture_info->lr_type[1] || picture_info->lr_type[2]) { const uint8_t unit_shift = 6 + picture->picture_parameter.loop_restoration_fields.lr_unit_shift; picture->picture_parameter.lr_unit_size[0] = (1 << unit_shift); picture->picture_parameter.lr_unit_size[1] = 1 << (unit_shift - picture->picture_parameter.loop_restoration_fields.lr_uv_shift); picture->picture_parameter.lr_unit_size[2] = picture->picture_parameter.lr_unit_size[1]; } else { for (i = 0; i < ARRAY_SIZE(picture->picture_parameter.lr_unit_size); ++i) { picture->picture_parameter.lr_unit_size[i] = (1 << 8); } } // Global Motion for (i = 0; i < AV1_REFS_PER_FRAME; ++i) { picture->picture_parameter.wm[i].invalid = picture_info->global_motion[i].invalid; picture->picture_parameter.wm[i].wmtype = picture_info->global_motion[i].wmtype; // VDPAU only has 6 wmmat[] elements, whereas Gallium provides 8. for (j = 0; j < ARRAY_SIZE(picture_info->global_motion[0].wmmat); ++j) { picture->picture_parameter.wm[i].wmmat[j] = picture_info->global_motion[i].wmmat[j]; } } picture->picture_parameter.matrix_coefficients = 0; // Tile Information picture->slice_parameter.slice_count = picture_info->num_tile_rows * picture_info->num_tile_cols; for (i = 0; i < picture->slice_parameter.slice_count; ++i) { const uint32_t start_offset = picture_info->tile_info[i * 2]; picture->slice_parameter.slice_data_offset[i] = start_offset; picture->slice_parameter.slice_data_size[i] = picture_info->tile_info[i * 2 + 1] - start_offset; } return VDP_STATUS_OK; } static void vlVdpDecoderFixVC1Startcode(uint32_t *num_buffers, const void *buffers[], unsigned sizes[]) { static const uint8_t vc1_startcode[] = { 0x00, 0x00, 0x01, 0x0D }; struct vl_vlc vlc = {}; unsigned i; /* search the first 64 bytes for a startcode */ vl_vlc_init(&vlc, *num_buffers, buffers, sizes); while (vl_vlc_search_byte(&vlc, 64*8, 0x00) && vl_vlc_bits_left(&vlc) >= 32) { uint32_t value = vl_vlc_peekbits(&vlc, 32); if (value == 0x0000010D || value == 0x0000010C || value == 0x0000010B) return; vl_vlc_eatbits(&vlc, 8); } /* none found, ok add one manually */ VDPAU_MSG(VDPAU_TRACE, "[VDPAU] Manually adding VC-1 startcode\n"); for (i = *num_buffers; i > 0; --i) { buffers[i] = buffers[i - 1]; sizes[i] = sizes[i - 1]; } ++(*num_buffers); buffers[0] = vc1_startcode; sizes[0] = 4; } static bool vlVdpQueryInterlacedH264(struct pipe_h264_picture_desc *h264) { if (h264->pps->sps->frame_mbs_only_flag) return false; return h264->field_pic_flag || /* PAFF */ h264->pps->sps->mb_adaptive_frame_field_flag; /* MBAFF */ } /** * Decode a compressed field/frame and render the result into a VdpVideoSurface. */ VdpStatus vlVdpDecoderRender(VdpDecoder decoder, VdpVideoSurface target, VdpPictureInfo const *picture_info, uint32_t bitstream_buffer_count, VdpBitstreamBuffer const *bitstream_buffers) { const void * buffers[bitstream_buffer_count + 1]; unsigned sizes[bitstream_buffer_count + 1]; vlVdpDecoder *vldecoder; vlVdpSurface *vlsurf; VdpStatus ret; struct pipe_screen *screen; struct pipe_video_codec *dec; bool buffer_support[2]; unsigned i; struct pipe_h264_sps sps_h264 = {}; struct pipe_h264_pps pps_h264 = { &sps_h264 }; struct pipe_h265_sps sps_h265 = {}; struct pipe_h265_pps pps_h265 = { &sps_h265 }; union { struct pipe_picture_desc base; struct pipe_mpeg12_picture_desc mpeg12; struct pipe_mpeg4_picture_desc mpeg4; struct pipe_vc1_picture_desc vc1; struct pipe_h264_picture_desc h264; struct pipe_h265_picture_desc h265; struct pipe_av1_picture_desc av1; } desc; bool picture_interlaced = false; if (!(picture_info && bitstream_buffers)) return VDP_STATUS_INVALID_POINTER; vldecoder = (vlVdpDecoder *)vlGetDataHTAB(decoder); if (!vldecoder) return VDP_STATUS_INVALID_HANDLE; dec = vldecoder->decoder; screen = dec->context->screen; vlsurf = (vlVdpSurface *)vlGetDataHTAB(target); if (!vlsurf) return VDP_STATUS_INVALID_HANDLE; if (vlsurf->device != vldecoder->device) return VDP_STATUS_HANDLE_DEVICE_MISMATCH; if (vlsurf->video_buffer != NULL && pipe_format_to_chroma_format(vlsurf->video_buffer->buffer_format) != dec->chroma_format) // TODO: Recreate decoder with correct chroma return VDP_STATUS_INVALID_CHROMA_TYPE; for (i = 0; i < bitstream_buffer_count; ++i) { buffers[i] = bitstream_buffers[i].bitstream; sizes[i] = bitstream_buffers[i].bitstream_bytes; } memset(&desc, 0, sizeof(desc)); desc.base.profile = dec->profile; switch (u_reduce_video_profile(dec->profile)) { case PIPE_VIDEO_FORMAT_MPEG12: ret = vlVdpDecoderRenderMpeg12(&desc.mpeg12, (VdpPictureInfoMPEG1Or2 *)picture_info); break; case PIPE_VIDEO_FORMAT_MPEG4: ret = vlVdpDecoderRenderMpeg4(&desc.mpeg4, (VdpPictureInfoMPEG4Part2 *)picture_info); break; case PIPE_VIDEO_FORMAT_VC1: if (dec->profile == PIPE_VIDEO_PROFILE_VC1_ADVANCED) vlVdpDecoderFixVC1Startcode(&bitstream_buffer_count, buffers, sizes); ret = vlVdpDecoderRenderVC1(&desc.vc1, (VdpPictureInfoVC1 *)picture_info); break; case PIPE_VIDEO_FORMAT_MPEG4_AVC: desc.h264.pps = &pps_h264; ret = vlVdpDecoderRenderH264(&desc.h264, (VdpPictureInfoH264 *)picture_info, dec->level); picture_interlaced = vlVdpQueryInterlacedH264(&desc.h264); break; case PIPE_VIDEO_FORMAT_HEVC: desc.h265.pps = &pps_h265; ret = vlVdpDecoderRenderH265(&desc.h265, (VdpPictureInfoHEVC *)picture_info); break; case PIPE_VIDEO_FORMAT_AV1: ret = vlVdpDecoderRenderAV1(&desc.av1, target, (VdpPictureInfoAV1 *)picture_info); break; default: return VDP_STATUS_INVALID_DECODER_PROFILE; } if (ret != VDP_STATUS_OK) return ret; buffer_support[0] = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM, PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE); buffer_support[1] = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM, PIPE_VIDEO_CAP_SUPPORTS_INTERLACED); if (vlsurf->video_buffer == NULL || !screen->is_video_format_supported(screen, vlsurf->video_buffer->buffer_format, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM) || !buffer_support[vlsurf->video_buffer->interlaced] || (picture_interlaced && !vlsurf->video_buffer->interlaced && buffer_support[1])) { mtx_lock(&vlsurf->device->mutex); /* destroy the old one */ if (vlsurf->video_buffer) vlsurf->video_buffer->destroy(vlsurf->video_buffer); /* set the buffer format to the prefered one */ vlsurf->templat.buffer_format = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM, PIPE_VIDEO_CAP_PREFERED_FORMAT); /* also set interlacing to decoders preferences */ vlsurf->templat.interlaced = screen->get_video_param(screen, dec->profile, PIPE_VIDEO_ENTRYPOINT_BITSTREAM, PIPE_VIDEO_CAP_PREFERS_INTERLACED) || picture_interlaced; /* and recreate the video buffer */ vlsurf->video_buffer = dec->context->create_video_buffer(dec->context, &vlsurf->templat); /* still no luck? get me out of here... */ if (!vlsurf->video_buffer) { mtx_unlock(&vlsurf->device->mutex); return VDP_STATUS_NO_IMPLEMENTATION; } vlVdpVideoSurfaceClear(vlsurf); mtx_unlock(&vlsurf->device->mutex); } mtx_lock(&vldecoder->mutex); dec->begin_frame(dec, vlsurf->video_buffer, &desc.base); dec->decode_bitstream(dec, vlsurf->video_buffer, &desc.base, bitstream_buffer_count, buffers, sizes); dec->end_frame(dec, vlsurf->video_buffer, &desc.base); mtx_unlock(&vldecoder->mutex); return ret; }