mirror of
https://gitlab.freedesktop.org/mesa/mesa.git
synced 2025-12-24 15:20:10 +01:00
Reviewed-by: Giancarlo Devich <gdevich@microsoft.com> Part-of: <https://gitlab.freedesktop.org/mesa/mesa/-/merge_requests/18944>
1286 lines
65 KiB
C++
1286 lines
65 KiB
C++
/*
|
||
* Copyright © Microsoft Corporation
|
||
*
|
||
* Permission is hereby granted, free of charge, to any person obtaining a
|
||
* copy of this software and associated documentation files (the "Software"),
|
||
* to deal in the Software without restriction, including without limitation
|
||
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
||
* and/or sell copies of the Software, and to permit persons to whom the
|
||
* Software is furnished to do so, subject to the following conditions:
|
||
*
|
||
* The above copyright notice and this permission notice (including the next
|
||
* paragraph) shall be included in all copies or substantial portions of the
|
||
* Software.
|
||
*
|
||
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
||
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
||
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
||
* IN THE SOFTWARE.
|
||
*/
|
||
|
||
#include "d3d12_screen.h"
|
||
#include "d3d12_video_screen.h"
|
||
#include "d3d12_format.h"
|
||
#include "util/u_video.h"
|
||
#include <directx/d3d12video.h>
|
||
#include <cmath>
|
||
|
||
#include <wrl/client.h>
|
||
using Microsoft::WRL::ComPtr;
|
||
|
||
#include "d3d12_video_types.h"
|
||
|
||
struct d3d12_encode_codec_support {
|
||
enum pipe_video_profile profile;
|
||
union {
|
||
struct {
|
||
enum pipe_h265_enc_pred_direction prediction_direction;
|
||
union pipe_h265_enc_cap_features hevc_features;
|
||
union pipe_h265_enc_cap_block_sizes hevc_block_sizes;
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC d3d12_caps;
|
||
} hevc_support;
|
||
// Can add more codecs for each codec specific caps here, for example:
|
||
// struct {
|
||
// D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_H264;
|
||
// } h264_support;
|
||
};
|
||
};
|
||
|
||
struct d3d12_video_resolution_to_level_mapping_entry
|
||
{
|
||
D3D12_VIDEO_ENCODER_PICTURE_RESOLUTION_DESC resolution;
|
||
uint32_t level;
|
||
};
|
||
|
||
static d3d12_video_resolution_to_level_mapping_entry
|
||
get_max_level_resolution_video_decode_support(D3D12_VIDEO_DECODE_CONFIGURATION decoderConfig,
|
||
DXGI_FORMAT format,
|
||
struct pipe_screen *pscreen,
|
||
bool &outSupportAny,
|
||
D3D12_FEATURE_DATA_VIDEO_DECODE_SUPPORT &outSupportedConfig)
|
||
{
|
||
d3d12_video_resolution_to_level_mapping_entry supportedResult = {};
|
||
outSupportAny = false;
|
||
outSupportedConfig = {};
|
||
|
||
ComPtr<ID3D12VideoDevice> spD3D12VideoDevice;
|
||
struct d3d12_screen *pD3D12Screen = (struct d3d12_screen *) pscreen;
|
||
if (FAILED(pD3D12Screen->dev->QueryInterface(IID_PPV_ARGS(spD3D12VideoDevice.GetAddressOf())))) {
|
||
// No video support in underlying d3d12 device (decode needs ID3D12VideoDevice)
|
||
return supportedResult;
|
||
}
|
||
|
||
d3d12_video_resolution_to_level_mapping_entry resolutionsLevelList[] = {
|
||
{ { 8192, 4320 }, 61 }, // 8k
|
||
{ { 7680, 4800 }, 61 }, // 8k - alternative
|
||
{ { 7680, 4320 }, 61 }, // 8k - alternative
|
||
{ { 4096, 2304 }, 52 }, // 2160p (4K)
|
||
{ { 4096, 2160 }, 52 }, // 2160p (4K) - alternative
|
||
{ { 2560, 1440 }, 51 }, // 1440p
|
||
{ { 1920, 1200 }, 5 }, // 1200p
|
||
{ { 1920, 1080 }, 42 }, // 1080p
|
||
{ { 1280, 720 }, 4 }, // 720p
|
||
{ { 800, 600 }, 31 },
|
||
};
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_DECODE_SUPPORT decodeSupport = {};
|
||
decodeSupport.Configuration = decoderConfig;
|
||
decodeSupport.DecodeFormat = format;
|
||
|
||
uint32_t idxResol = 0;
|
||
while ((idxResol < ARRAY_SIZE(resolutionsLevelList)) && !outSupportAny) {
|
||
|
||
decodeSupport.Width = resolutionsLevelList[idxResol].resolution.Width;
|
||
decodeSupport.Height = resolutionsLevelList[idxResol].resolution.Height;
|
||
|
||
if (SUCCEEDED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_DECODE_SUPPORT,
|
||
&decodeSupport,
|
||
sizeof(decodeSupport)))) {
|
||
|
||
if (((decodeSupport.SupportFlags & D3D12_VIDEO_DECODE_SUPPORT_FLAG_SUPPORTED) != 0) ||
|
||
decodeSupport.DecodeTier > D3D12_VIDEO_DECODE_TIER_NOT_SUPPORTED) {
|
||
|
||
outSupportAny = true;
|
||
outSupportedConfig = decodeSupport;
|
||
supportedResult = resolutionsLevelList[idxResol];
|
||
}
|
||
}
|
||
|
||
idxResol++;
|
||
}
|
||
|
||
return supportedResult;
|
||
}
|
||
|
||
static bool
|
||
d3d12_has_video_decode_support(struct pipe_screen *pscreen, enum pipe_video_profile profile)
|
||
{
|
||
ComPtr<ID3D12VideoDevice> spD3D12VideoDevice;
|
||
struct d3d12_screen *pD3D12Screen = (struct d3d12_screen *) pscreen;
|
||
if (FAILED(pD3D12Screen->dev->QueryInterface(IID_PPV_ARGS(spD3D12VideoDevice.GetAddressOf())))) {
|
||
// No video support in underlying d3d12 device (needs ID3D12VideoDevice)
|
||
return 0;
|
||
}
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_FEATURE_AREA_SUPPORT VideoFeatureAreaSupport = {};
|
||
if (FAILED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_FEATURE_AREA_SUPPORT,
|
||
&VideoFeatureAreaSupport,
|
||
sizeof(VideoFeatureAreaSupport)))) {
|
||
return false;
|
||
}
|
||
|
||
// Supported profiles below
|
||
bool supportsProfile = false;
|
||
switch (profile) {
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_BASELINE:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_EXTENDED:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_CONSTRAINED_BASELINE:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_MAIN:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_HIGH:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_HIGH10:
|
||
case PIPE_VIDEO_PROFILE_HEVC_MAIN:
|
||
case PIPE_VIDEO_PROFILE_HEVC_MAIN_10:
|
||
case PIPE_VIDEO_PROFILE_AV1_MAIN:
|
||
case PIPE_VIDEO_PROFILE_VP9_PROFILE0:
|
||
case PIPE_VIDEO_PROFILE_VP9_PROFILE2:
|
||
{
|
||
supportsProfile = true;
|
||
} break;
|
||
default:
|
||
supportsProfile = false;
|
||
}
|
||
|
||
return VideoFeatureAreaSupport.VideoDecodeSupport && supportsProfile;
|
||
}
|
||
|
||
static bool
|
||
d3d12_video_encode_max_supported_level_for_profile(const D3D12_VIDEO_ENCODER_CODEC &argCodec,
|
||
const D3D12_VIDEO_ENCODER_PROFILE_DESC &argTargetProfile,
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING &minLvl,
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING &maxLvl,
|
||
ID3D12VideoDevice3 *pD3D12VideoDevice)
|
||
{
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_PROFILE_LEVEL capLevelData = {};
|
||
capLevelData.NodeIndex = 0;
|
||
capLevelData.Codec = argCodec;
|
||
capLevelData.Profile = argTargetProfile;
|
||
capLevelData.MinSupportedLevel = minLvl;
|
||
capLevelData.MaxSupportedLevel = maxLvl;
|
||
|
||
if (FAILED(pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_PROFILE_LEVEL,
|
||
&capLevelData,
|
||
sizeof(capLevelData)))) {
|
||
return false;
|
||
}
|
||
|
||
return capLevelData.IsSupported;
|
||
}
|
||
|
||
static bool
|
||
d3d12_video_encode_max_supported_resolution(const D3D12_VIDEO_ENCODER_CODEC &argTargetCodec,
|
||
D3D12_VIDEO_ENCODER_PICTURE_RESOLUTION_DESC &maxResolution,
|
||
ID3D12VideoDevice3 *pD3D12VideoDevice)
|
||
{
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_OUTPUT_RESOLUTION_RATIOS_COUNT capResRatiosCountData = { 0, argTargetCodec, 0 };
|
||
|
||
if (FAILED(pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_OUTPUT_RESOLUTION_RATIOS_COUNT,
|
||
&capResRatiosCountData,
|
||
sizeof(capResRatiosCountData)))) {
|
||
return false;
|
||
}
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_OUTPUT_RESOLUTION capOutputResolutionData = {};
|
||
capOutputResolutionData.NodeIndex = 0;
|
||
capOutputResolutionData.Codec = argTargetCodec;
|
||
capOutputResolutionData.ResolutionRatiosCount = capResRatiosCountData.ResolutionRatiosCount;
|
||
|
||
std::vector<D3D12_VIDEO_ENCODER_PICTURE_RESOLUTION_RATIO_DESC> ratiosTmpOutput;
|
||
if (capResRatiosCountData.ResolutionRatiosCount > 0) {
|
||
ratiosTmpOutput.resize(capResRatiosCountData.ResolutionRatiosCount);
|
||
capOutputResolutionData.pResolutionRatios = ratiosTmpOutput.data();
|
||
} else {
|
||
capOutputResolutionData.pResolutionRatios = nullptr;
|
||
}
|
||
|
||
if (FAILED(pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_OUTPUT_RESOLUTION,
|
||
&capOutputResolutionData,
|
||
sizeof(capOutputResolutionData))) ||
|
||
!capOutputResolutionData.IsSupported) {
|
||
return false;
|
||
}
|
||
|
||
maxResolution = capOutputResolutionData.MaxResolutionSupported;
|
||
|
||
return true;
|
||
}
|
||
|
||
static uint32_t
|
||
d3d12_video_encode_supported_references_per_frame_structures(const D3D12_VIDEO_ENCODER_CODEC &codec,
|
||
D3D12_VIDEO_ENCODER_PROFILE_DESC profile,
|
||
ID3D12VideoDevice3 *pD3D12VideoDevice)
|
||
{
|
||
uint32_t supportedMaxRefFrames = 0u;
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT capPictureControlData = {};
|
||
capPictureControlData.NodeIndex = 0;
|
||
capPictureControlData.Codec = codec;
|
||
|
||
if(codec == D3D12_VIDEO_ENCODER_CODEC_H264) {
|
||
D3D12_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT_H264 h264PictureControl = {};
|
||
capPictureControlData.Profile = profile;
|
||
capPictureControlData.PictureSupport.pH264Support = &h264PictureControl;
|
||
capPictureControlData.PictureSupport.DataSize = sizeof(h264PictureControl);
|
||
HRESULT hr = pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT,
|
||
&capPictureControlData,
|
||
sizeof(capPictureControlData));
|
||
if (FAILED(hr)) {
|
||
debug_printf("CheckFeatureSupport failed with HR %x\n", hr);
|
||
}
|
||
|
||
if (capPictureControlData.IsSupported) {
|
||
/* This attribute determines the maximum number of reference
|
||
* frames supported for encoding.
|
||
*
|
||
* Note: for H.264 encoding, the value represents the maximum number
|
||
* of reference frames for both the reference picture list 0 (bottom
|
||
* 16 bits) and the reference picture list 1 (top 16 bits).
|
||
*/
|
||
uint32_t maxRefForL0 = std::min(capPictureControlData.PictureSupport.pH264Support->MaxL0ReferencesForP,
|
||
capPictureControlData.PictureSupport.pH264Support->MaxL0ReferencesForB);
|
||
uint32_t maxRefForL1 = capPictureControlData.PictureSupport.pH264Support->MaxL1ReferencesForB;
|
||
supportedMaxRefFrames = (maxRefForL0 & 0xffff) | ((maxRefForL1 & 0xffff) << 16);
|
||
}
|
||
} else if(codec == D3D12_VIDEO_ENCODER_CODEC_HEVC) {
|
||
D3D12_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT_HEVC hevcPictureControl = {};
|
||
capPictureControlData.Profile = profile;
|
||
capPictureControlData.PictureSupport.pHEVCSupport = &hevcPictureControl;
|
||
capPictureControlData.PictureSupport.DataSize = sizeof(hevcPictureControl);
|
||
HRESULT hr = pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_CODEC_PICTURE_CONTROL_SUPPORT,
|
||
&capPictureControlData,
|
||
sizeof(capPictureControlData));
|
||
if (FAILED(hr)) {
|
||
debug_printf("CheckFeatureSupport failed with HR %x\n", hr);
|
||
}
|
||
|
||
if (capPictureControlData.IsSupported) {
|
||
/* This attribute determines the maximum number of reference
|
||
* frames supported for encoding.
|
||
*
|
||
* Note: for H.265 encoding, the value represents the maximum number
|
||
* of reference frames for both the reference picture list 0 (bottom
|
||
* 16 bits) and the reference picture list 1 (top 16 bits).
|
||
*/
|
||
uint32_t maxRefForL0 = std::min(capPictureControlData.PictureSupport.pHEVCSupport->MaxL0ReferencesForP,
|
||
capPictureControlData.PictureSupport.pHEVCSupport->MaxL0ReferencesForB);
|
||
uint32_t maxRefForL1 = capPictureControlData.PictureSupport.pHEVCSupport->MaxL1ReferencesForB;
|
||
supportedMaxRefFrames = (maxRefForL0 & 0xffff) | ((maxRefForL1 & 0xffff) << 16);
|
||
}
|
||
}
|
||
|
||
return supportedMaxRefFrames;
|
||
}
|
||
|
||
static uint32_t
|
||
d3d12_video_encode_supported_slice_structures(const D3D12_VIDEO_ENCODER_CODEC &codec,
|
||
D3D12_VIDEO_ENCODER_PROFILE_DESC profile,
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING level,
|
||
ID3D12VideoDevice3 *pD3D12VideoDevice)
|
||
{
|
||
uint32_t supportedSliceStructuresBitMask = PIPE_VIDEO_CAP_SLICE_STRUCTURE_NONE;
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE capDataSubregionLayout = {};
|
||
capDataSubregionLayout.NodeIndex = 0;
|
||
capDataSubregionLayout.Codec = codec;
|
||
capDataSubregionLayout.Profile = profile;
|
||
capDataSubregionLayout.Level = level;
|
||
|
||
/**
|
||
* pipe_video_cap_slice_structure
|
||
*
|
||
* This attribute determines slice structures supported by the
|
||
* driver for encoding. This attribute is a hint to the user so
|
||
* that he can choose a suitable surface size and how to arrange
|
||
* the encoding process of multiple slices per frame.
|
||
*
|
||
* More specifically, for H.264 encoding, this attribute
|
||
* determines the range of accepted values to
|
||
* h264_slice_descriptor::macroblock_address and
|
||
* h264_slice_descriptor::num_macroblocks.
|
||
*
|
||
* For HEVC, similarly determines the ranges for
|
||
* slice_segment_address
|
||
* num_ctu_in_slice
|
||
*/
|
||
capDataSubregionLayout.SubregionMode =
|
||
D3D12_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE_UNIFORM_PARTITIONING_SUBREGIONS_PER_FRAME;
|
||
HRESULT hr = pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE,
|
||
&capDataSubregionLayout,
|
||
sizeof(capDataSubregionLayout));
|
||
if (FAILED(hr)) {
|
||
debug_printf("CheckFeatureSupport failed with HR %x\n", hr);
|
||
} else if (capDataSubregionLayout.IsSupported) {
|
||
/* This would be setting N subregions per frame in this D3D12 mode where N = (height/blocksize) / K */
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_EQUAL_MULTI_ROWS;
|
||
/* Assuming height/blocksize >= max_supported_slices, which is reported
|
||
in PIPE_VIDEO_CAP_ENC_MAX_SLICES_PER_FRAME and should be checked by the client*/
|
||
/* This would be setting N subregions per frame in this D3D12 mode where N = (height/blocksize) */
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_EQUAL_ROWS;
|
||
/* This is ok, would be setting K rows per subregions in this D3D12 mode (and rounding the last one) */
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_POWER_OF_TWO_ROWS;
|
||
}
|
||
|
||
capDataSubregionLayout.SubregionMode =
|
||
D3D12_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE_UNIFORM_PARTITIONING_ROWS_PER_SUBREGION;
|
||
hr = pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE,
|
||
&capDataSubregionLayout,
|
||
sizeof(capDataSubregionLayout));
|
||
if (FAILED(hr)) {
|
||
debug_printf("CheckFeatureSupport failed with HR %x\n", hr);
|
||
} else if (capDataSubregionLayout.IsSupported) {
|
||
/* This would be setting K rows per subregions in this D3D12 mode */
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_EQUAL_MULTI_ROWS;
|
||
/* Assuming height/blocksize >= max_supported_slices, which is reported
|
||
in PIPE_VIDEO_CAP_ENC_MAX_SLICES_PER_FRAME and should be checked by the client*/
|
||
/* This would be setting 1 row per subregion in this D3D12 mode */
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_EQUAL_ROWS;
|
||
/* This is ok, would be setting K rows per subregions in this D3D12 mode (and rounding the last one) */
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_POWER_OF_TWO_ROWS;
|
||
}
|
||
|
||
/* Needs more work in VA frontend to support VAEncMiscParameterMaxSliceSize
|
||
and the driver potentially reporting back status in VACodedBufferSegment */
|
||
|
||
/*capDataSubregionLayout.SubregionMode = D3D12_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE_BYTES_PER_SUBREGION;
|
||
hr = pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE,
|
||
&capDataSubregionLayout,
|
||
sizeof(capDataSubregionLayout));
|
||
if (FAILED(hr)) {
|
||
debug_printf("CheckFeatureSupport failed with HR %x\n", hr);
|
||
} else if (capDataSubregionLayout.IsSupported) {
|
||
supportedSliceStructuresBitMask |= PIPE_VIDEO_CAP_SLICE_STRUCTURE_MAX_SLICE_SIZE;
|
||
}*/
|
||
|
||
return supportedSliceStructuresBitMask;
|
||
}
|
||
|
||
static bool
|
||
d3d12_video_encode_max_supported_slices(const D3D12_VIDEO_ENCODER_CODEC &argTargetCodec,
|
||
D3D12_VIDEO_ENCODER_PICTURE_RESOLUTION_DESC maxResolution,
|
||
DXGI_FORMAT encodeFormat,
|
||
uint32_t &outMaxSlices,
|
||
ID3D12VideoDevice3 *pD3D12VideoDevice,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT codecSupport)
|
||
{
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_SUPPORT capEncoderSupportData = {};
|
||
capEncoderSupportData.NodeIndex = 0;
|
||
capEncoderSupportData.Codec = argTargetCodec;
|
||
capEncoderSupportData.InputFormat = encodeFormat;
|
||
capEncoderSupportData.RateControl = {};
|
||
capEncoderSupportData.RateControl.Mode = D3D12_VIDEO_ENCODER_RATE_CONTROL_MODE_CQP;
|
||
capEncoderSupportData.RateControl.TargetFrameRate.Numerator = 60;
|
||
capEncoderSupportData.RateControl.TargetFrameRate.Denominator = 1;
|
||
D3D12_VIDEO_ENCODER_RATE_CONTROL_CQP rcCqp = { 25, 25, 25 };
|
||
capEncoderSupportData.RateControl.ConfigParams.pConfiguration_CQP = &rcCqp;
|
||
capEncoderSupportData.RateControl.ConfigParams.DataSize = sizeof(rcCqp);
|
||
capEncoderSupportData.IntraRefresh = D3D12_VIDEO_ENCODER_INTRA_REFRESH_MODE_NONE;
|
||
capEncoderSupportData.ResolutionsListCount = 1;
|
||
capEncoderSupportData.pResolutionList = &maxResolution;
|
||
capEncoderSupportData.MaxReferenceFramesInDPB = 1;
|
||
capEncoderSupportData.SubregionFrameEncoding =
|
||
D3D12_VIDEO_ENCODER_FRAME_SUBREGION_LAYOUT_MODE_UNIFORM_PARTITIONING_SUBREGIONS_PER_FRAME;
|
||
|
||
/*
|
||
All codec structures must be declared outside the switch statement to be
|
||
present in memory (stack scope) when calling CheckFeatureSupport below
|
||
*/
|
||
D3D12_VIDEO_ENCODER_PROFILE_H264 h264prof = {};
|
||
D3D12_VIDEO_ENCODER_LEVELS_H264 h264lvl = {};
|
||
D3D12_VIDEO_ENCODER_SEQUENCE_GOP_STRUCTURE_H264 h264Gop = { 1, 0, 0, 0, 0 };
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_H264 h264Config = {};
|
||
D3D12_VIDEO_ENCODER_PROFILE_HEVC hevcprof = D3D12_VIDEO_ENCODER_PROFILE_HEVC_MAIN;
|
||
D3D12_VIDEO_ENCODER_LEVEL_TIER_CONSTRAINTS_HEVC hevcLvl = { D3D12_VIDEO_ENCODER_LEVELS_HEVC_62, D3D12_VIDEO_ENCODER_TIER_HEVC_HIGH };
|
||
D3D12_VIDEO_ENCODER_SEQUENCE_GOP_STRUCTURE_HEVC hevcGop = { 1, 0, 0 };
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC hevcConfig = {};
|
||
switch (argTargetCodec) {
|
||
case D3D12_VIDEO_ENCODER_CODEC_H264:
|
||
{
|
||
// assert(codecSupport.pH264Support); // Fill this in caller if ever used
|
||
capEncoderSupportData.SuggestedProfile.pH264Profile = &h264prof;
|
||
capEncoderSupportData.SuggestedProfile.DataSize = sizeof(h264prof);
|
||
capEncoderSupportData.SuggestedLevel.pH264LevelSetting = &h264lvl;
|
||
capEncoderSupportData.SuggestedLevel.DataSize = sizeof(h264lvl);
|
||
capEncoderSupportData.CodecGopSequence.pH264GroupOfPictures = &h264Gop;
|
||
capEncoderSupportData.CodecGopSequence.DataSize = sizeof(h264Gop);
|
||
capEncoderSupportData.CodecConfiguration.DataSize = sizeof(h264Config);
|
||
capEncoderSupportData.CodecConfiguration.pH264Config = &h264Config;
|
||
} break;
|
||
|
||
case D3D12_VIDEO_ENCODER_CODEC_HEVC:
|
||
{
|
||
/* Only read from codecSupport.pHEVCSupport in this case (union of pointers definition) */
|
||
assert(codecSupport.pHEVCSupport);
|
||
hevcConfig = {
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_FLAG_NONE,
|
||
codecSupport.pHEVCSupport->MinLumaCodingUnitSize,
|
||
codecSupport.pHEVCSupport->MaxLumaCodingUnitSize,
|
||
codecSupport.pHEVCSupport->MinLumaTransformUnitSize,
|
||
codecSupport.pHEVCSupport->MaxLumaTransformUnitSize,
|
||
codecSupport.pHEVCSupport->max_transform_hierarchy_depth_inter,
|
||
codecSupport.pHEVCSupport->max_transform_hierarchy_depth_intra,
|
||
};
|
||
capEncoderSupportData.SuggestedProfile.pHEVCProfile = &hevcprof;
|
||
capEncoderSupportData.SuggestedProfile.DataSize = sizeof(hevcprof);
|
||
capEncoderSupportData.SuggestedLevel.pHEVCLevelSetting = &hevcLvl;
|
||
capEncoderSupportData.SuggestedLevel.DataSize = sizeof(hevcLvl);
|
||
capEncoderSupportData.CodecGopSequence.pHEVCGroupOfPictures = &hevcGop;
|
||
capEncoderSupportData.CodecGopSequence.DataSize = sizeof(hevcGop);
|
||
capEncoderSupportData.CodecConfiguration.DataSize = sizeof(hevcConfig);
|
||
capEncoderSupportData.CodecConfiguration.pHEVCConfig = &hevcConfig;
|
||
} break;
|
||
|
||
default:
|
||
{
|
||
unreachable("Unsupported D3D12_VIDEO_ENCODER_CODEC");
|
||
} break;
|
||
}
|
||
|
||
// prepare inout storage for the resolution dependent result.
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_RESOLUTION_SUPPORT_LIMITS resolutionDepCaps = {};
|
||
capEncoderSupportData.pResolutionDependentSupport = &resolutionDepCaps;
|
||
|
||
HRESULT hr = pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_SUPPORT,
|
||
&capEncoderSupportData,
|
||
sizeof(capEncoderSupportData));
|
||
if (FAILED(hr)) {
|
||
debug_printf("CheckFeatureSupport failed with HR %x\n", hr);
|
||
return false;
|
||
} else {
|
||
bool configSupported =
|
||
(((capEncoderSupportData.SupportFlags & D3D12_VIDEO_ENCODER_SUPPORT_FLAG_GENERAL_SUPPORT_OK) != 0) &&
|
||
(capEncoderSupportData.ValidationFlags == D3D12_VIDEO_ENCODER_VALIDATION_FLAG_NONE));
|
||
|
||
outMaxSlices = resolutionDepCaps.MaxSubregionsNumber;
|
||
return configSupported;
|
||
}
|
||
}
|
||
|
||
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC
|
||
static d3d12_video_encode_get_hevc_codec_support ( const D3D12_VIDEO_ENCODER_CODEC &argCodec,
|
||
const D3D12_VIDEO_ENCODER_PROFILE_DESC &argTargetProfile,
|
||
ID3D12VideoDevice3 *pD3D12VideoDevice)
|
||
{
|
||
|
||
constexpr unsigned c_hevcConfigurationSets = 5u;
|
||
const D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC hevcConfigurationSets[c_hevcConfigurationSets] =
|
||
{
|
||
{
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_NONE,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_8x8,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_32x32,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_4x4,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_32x32,
|
||
3u,
|
||
3u,
|
||
},
|
||
{
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_NONE,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_8x8,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_32x32,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_4x4,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_32x32,
|
||
0u,
|
||
0u,
|
||
},
|
||
{
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_NONE,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_8x8,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_32x32,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_4x4,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_32x32,
|
||
2u,
|
||
2u,
|
||
},
|
||
{
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_NONE,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_8x8,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_64x64,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_4x4,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_32x32,
|
||
2u,
|
||
2u,
|
||
},
|
||
{
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_NONE,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_8x8,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_CUSIZE_64x64,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_4x4,
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_HEVC_TUSIZE_32x32,
|
||
4u,
|
||
4u,
|
||
},
|
||
};
|
||
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC hevcCodecCaps = { };
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT capCodecConfigData = { };
|
||
capCodecConfigData.NodeIndex = 0;
|
||
capCodecConfigData.Codec = D3D12_VIDEO_ENCODER_CODEC_HEVC;
|
||
capCodecConfigData.Profile = argTargetProfile;
|
||
capCodecConfigData.CodecSupportLimits.pHEVCSupport = &hevcCodecCaps;
|
||
capCodecConfigData.CodecSupportLimits.DataSize = sizeof(hevcCodecCaps);
|
||
|
||
for (auto hevc_config : hevcConfigurationSets) {
|
||
hevcCodecCaps = hevc_config;
|
||
if(SUCCEEDED(pD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT, &capCodecConfigData, sizeof(capCodecConfigData))
|
||
&& capCodecConfigData.IsSupported)) {
|
||
hevc_config.SupportFlags = hevcCodecCaps.SupportFlags;
|
||
return hevc_config;
|
||
}
|
||
}
|
||
|
||
/* If we reach this point, the underlying HW/Driver might need a new configuration
|
||
added to the table and be iterated above */
|
||
unreachable("D3D12: Couldn't find HEVC supported configuration arguments.");
|
||
return hevcCodecCaps;
|
||
}
|
||
|
||
static bool
|
||
d3d12_has_video_encode_support(struct pipe_screen *pscreen,
|
||
enum pipe_video_profile profile,
|
||
uint32_t &maxLvlSpec,
|
||
D3D12_VIDEO_ENCODER_PICTURE_RESOLUTION_DESC &maxRes,
|
||
uint32_t &maxSlices,
|
||
uint32_t &supportedSliceStructures,
|
||
uint32_t &maxReferencesPerFrame,
|
||
struct d3d12_encode_codec_support& codecSupport)
|
||
{
|
||
ComPtr<ID3D12VideoDevice3> spD3D12VideoDevice;
|
||
struct d3d12_screen *pD3D12Screen = (struct d3d12_screen *) pscreen;
|
||
if (FAILED(pD3D12Screen->dev->QueryInterface(IID_PPV_ARGS(spD3D12VideoDevice.GetAddressOf())))) {
|
||
// No video encode support in underlying d3d12 device (needs ID3D12VideoDevice3)
|
||
return 0;
|
||
}
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_FEATURE_AREA_SUPPORT VideoFeatureAreaSupport = {};
|
||
if (FAILED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_FEATURE_AREA_SUPPORT,
|
||
&VideoFeatureAreaSupport,
|
||
sizeof(VideoFeatureAreaSupport)))) {
|
||
return false;
|
||
}
|
||
D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT d3d12_codec_support = { };
|
||
bool supportsProfile = false;
|
||
switch (profile) {
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_CONSTRAINED_BASELINE:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_BASELINE:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_MAIN:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_HIGH:
|
||
case PIPE_VIDEO_PROFILE_MPEG4_AVC_HIGH10:
|
||
{
|
||
D3D12_VIDEO_ENCODER_PROFILE_DESC profDesc = {};
|
||
D3D12_VIDEO_ENCODER_PROFILE_H264 profH264 =
|
||
d3d12_video_encoder_convert_profile_to_d3d12_enc_profile_h264(profile);
|
||
profDesc.DataSize = sizeof(profH264);
|
||
profDesc.pH264Profile = &profH264;
|
||
D3D12_VIDEO_ENCODER_CODEC codecDesc = d3d12_video_encoder_convert_codec_to_d3d12_enc_codec(profile);
|
||
D3D12_VIDEO_ENCODER_LEVELS_H264 minLvlSettingH264 = static_cast<D3D12_VIDEO_ENCODER_LEVELS_H264>(0);
|
||
D3D12_VIDEO_ENCODER_LEVELS_H264 maxLvlSettingH264 = static_cast<D3D12_VIDEO_ENCODER_LEVELS_H264>(0);
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING minLvl = {};
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING maxLvl = {};
|
||
minLvl.pH264LevelSetting = &minLvlSettingH264;
|
||
minLvl.DataSize = sizeof(minLvlSettingH264);
|
||
maxLvl.pH264LevelSetting = &maxLvlSettingH264;
|
||
maxLvl.DataSize = sizeof(maxLvlSettingH264);
|
||
if (d3d12_video_encode_max_supported_level_for_profile(codecDesc,
|
||
profDesc,
|
||
minLvl,
|
||
maxLvl,
|
||
spD3D12VideoDevice.Get())) {
|
||
uint32_t constraintset3flag = false;
|
||
d3d12_video_encoder_convert_from_d3d12_level_h264(maxLvlSettingH264, maxLvlSpec, constraintset3flag);
|
||
supportsProfile = true;
|
||
|
||
DXGI_FORMAT encodeFormat = d3d12_convert_pipe_video_profile_to_dxgi_format(profile);
|
||
supportsProfile = supportsProfile &&
|
||
d3d12_video_encode_max_supported_resolution(codecDesc, maxRes, spD3D12VideoDevice.Get());
|
||
|
||
D3D12_VIDEO_ENCODER_PROFILE_DESC profile;
|
||
profile.pH264Profile = &profH264;
|
||
profile.DataSize = sizeof(profH264);
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING level;
|
||
level.pH264LevelSetting = &maxLvlSettingH264;
|
||
level.DataSize = sizeof(maxLvlSettingH264);
|
||
supportedSliceStructures = d3d12_video_encode_supported_slice_structures(codecDesc,
|
||
profile,
|
||
level,
|
||
spD3D12VideoDevice.Get());
|
||
if (supportedSliceStructures == PIPE_VIDEO_CAP_SLICE_STRUCTURE_NONE)
|
||
maxSlices = 0;
|
||
else
|
||
supportsProfile = supportsProfile && d3d12_video_encode_max_supported_slices(codecDesc,
|
||
maxRes,
|
||
encodeFormat,
|
||
maxSlices,
|
||
spD3D12VideoDevice.Get(),
|
||
d3d12_codec_support);
|
||
maxReferencesPerFrame =
|
||
d3d12_video_encode_supported_references_per_frame_structures(codecDesc,
|
||
profile,
|
||
spD3D12VideoDevice.Get());
|
||
}
|
||
} break;
|
||
case PIPE_VIDEO_PROFILE_HEVC_MAIN:
|
||
case PIPE_VIDEO_PROFILE_HEVC_MAIN_10:
|
||
{
|
||
D3D12_VIDEO_ENCODER_PROFILE_DESC profDesc = {};
|
||
D3D12_VIDEO_ENCODER_PROFILE_HEVC profHEVC =
|
||
d3d12_video_encoder_convert_profile_to_d3d12_enc_profile_hevc(profile);
|
||
profDesc.DataSize = sizeof(profHEVC);
|
||
profDesc.pHEVCProfile = &profHEVC;
|
||
D3D12_VIDEO_ENCODER_CODEC codecDesc = d3d12_video_encoder_convert_codec_to_d3d12_enc_codec(profile);
|
||
D3D12_VIDEO_ENCODER_LEVEL_TIER_CONSTRAINTS_HEVC minLvlSettingHEVC = { };
|
||
D3D12_VIDEO_ENCODER_LEVEL_TIER_CONSTRAINTS_HEVC maxLvlSettingHEVC = { };
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING minLvl = {};
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING maxLvl = {};
|
||
minLvl.pHEVCLevelSetting = &minLvlSettingHEVC;
|
||
minLvl.DataSize = sizeof(minLvlSettingHEVC);
|
||
maxLvl.pHEVCLevelSetting = &maxLvlSettingHEVC;
|
||
maxLvl.DataSize = sizeof(maxLvlSettingHEVC);
|
||
if (d3d12_video_encode_max_supported_level_for_profile(codecDesc,
|
||
profDesc,
|
||
minLvl,
|
||
maxLvl,
|
||
spD3D12VideoDevice.Get())) {
|
||
d3d12_video_encoder_convert_from_d3d12_level_hevc(maxLvlSettingHEVC.Level, maxLvlSpec);
|
||
supportsProfile = true;
|
||
|
||
D3D12_VIDEO_ENCODER_PROFILE_DESC d3d12_profile;
|
||
d3d12_profile.pHEVCProfile = &profHEVC;
|
||
d3d12_profile.DataSize = sizeof(profHEVC);
|
||
D3D12_VIDEO_ENCODER_LEVEL_SETTING level;
|
||
level.pHEVCLevelSetting = &maxLvlSettingHEVC;
|
||
level.DataSize = sizeof(maxLvlSettingHEVC);
|
||
supportedSliceStructures = d3d12_video_encode_supported_slice_structures(codecDesc,
|
||
d3d12_profile,
|
||
level,
|
||
spD3D12VideoDevice.Get());
|
||
|
||
maxReferencesPerFrame =
|
||
d3d12_video_encode_supported_references_per_frame_structures(codecDesc,
|
||
d3d12_profile,
|
||
spD3D12VideoDevice.Get());
|
||
|
||
codecSupport.hevc_support.d3d12_caps = d3d12_video_encode_get_hevc_codec_support(codecDesc,
|
||
profDesc,
|
||
spD3D12VideoDevice.Get());
|
||
d3d12_codec_support.DataSize = sizeof(codecSupport.hevc_support.d3d12_caps);
|
||
d3d12_codec_support.pHEVCSupport = &codecSupport.hevc_support.d3d12_caps;
|
||
|
||
/* get_video_param sets pipe_features.bits.config_supported = 1
|
||
to distinguish between supported cap with all bits off and unsupported by driver
|
||
with value = 0
|
||
*/
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.config_supported = 1;
|
||
codecSupport.hevc_support.hevc_features.bits.config_supported = 1;
|
||
|
||
// Fill codecSupport.hevc_support
|
||
|
||
uint8_t minCuSize = d3d12_video_encoder_convert_12cusize_to_pixel_size_hevc(codecSupport.hevc_support.d3d12_caps.MinLumaCodingUnitSize);
|
||
uint8_t maxCuSize = d3d12_video_encoder_convert_12cusize_to_pixel_size_hevc(codecSupport.hevc_support.d3d12_caps.MaxLumaCodingUnitSize);
|
||
uint8_t MinCbLog2SizeY = std::log2(minCuSize);
|
||
uint8_t CtbLog2SizeY = std::log2(maxCuSize);
|
||
uint8_t minTuSize = d3d12_video_encoder_convert_12tusize_to_pixel_size_hevc(codecSupport.hevc_support.d3d12_caps.MinLumaTransformUnitSize);
|
||
uint8_t maxTuSize = d3d12_video_encoder_convert_12tusize_to_pixel_size_hevc(codecSupport.hevc_support.d3d12_caps.MaxLumaTransformUnitSize);
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_max_coding_tree_block_size_minus3
|
||
= static_cast<uint8_t>(CtbLog2SizeY - 3);
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_min_coding_tree_block_size_minus3
|
||
= static_cast<uint8_t>(CtbLog2SizeY - 3);
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_min_luma_coding_block_size_minus3
|
||
= static_cast<uint8_t>(MinCbLog2SizeY - 3);
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_max_luma_transform_block_size_minus2
|
||
= static_cast<uint8_t>(std::log2(maxTuSize) - 2);
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_min_luma_transform_block_size_minus2
|
||
= static_cast<uint8_t>(std::log2(minTuSize) - 2);
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.max_max_transform_hierarchy_depth_inter
|
||
= codecSupport.hevc_support.d3d12_caps.max_transform_hierarchy_depth_inter;
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.min_max_transform_hierarchy_depth_inter
|
||
= codecSupport.hevc_support.d3d12_caps.max_transform_hierarchy_depth_inter;
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.max_max_transform_hierarchy_depth_intra
|
||
= codecSupport.hevc_support.d3d12_caps.max_transform_hierarchy_depth_intra;
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.min_max_transform_hierarchy_depth_intra
|
||
= codecSupport.hevc_support.d3d12_caps.max_transform_hierarchy_depth_intra;
|
||
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_max_pcm_coding_block_size_minus3 = 0; // No PCM Supported
|
||
codecSupport.hevc_support.hevc_block_sizes.bits.log2_min_pcm_coding_block_size_minus3 = 0; // No PCM Supported
|
||
|
||
// Feature flags
|
||
|
||
uint32_t ref_l0 = maxReferencesPerFrame & 0xffff;
|
||
uint32_t ref_l1 = maxReferencesPerFrame >> 16 & 0xffff;
|
||
|
||
codecSupport.hevc_support.prediction_direction = PIPE_H265_PRED_DIRECTION_ALL;
|
||
if(ref_l0)
|
||
codecSupport.hevc_support.prediction_direction |= PIPE_H265_PRED_DIRECTION_PREVIOUS;
|
||
if(ref_l1)
|
||
codecSupport.hevc_support.prediction_direction |= PIPE_H265_PRED_DIRECTION_FUTURE;
|
||
|
||
codecSupport.hevc_support.hevc_features.bits.separate_colour_planes = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.scaling_lists = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.pcm = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.temporal_mvp = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.strong_intra_smoothing = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.dependent_slices = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.sign_data_hiding = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.weighted_prediction = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.transquant_bypass = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
codecSupport.hevc_support.hevc_features.bits.deblocking_filter_disable = PIPE_H265_ENC_FEATURE_NOT_SUPPORTED;
|
||
|
||
/* cu_qp_delta always required to be 1 in https://github.com/microsoft/DirectX-Specs/blob/master/d3d/D3D12VideoEncoding.md */
|
||
codecSupport.hevc_support.hevc_features.bits.cu_qp_delta = (PIPE_H265_ENC_FEATURE_SUPPORTED | PIPE_H265_ENC_FEATURE_REQUIRED);
|
||
|
||
if ((codecSupport.hevc_support.d3d12_caps.SupportFlags & D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_P_FRAMES_IMPLEMENTED_AS_LOW_DELAY_B_FRAMES) != 0)
|
||
codecSupport.hevc_support.prediction_direction |= PIPE_H265_PRED_DIRECTION_BI_NOT_EMPTY;
|
||
|
||
if ((codecSupport.hevc_support.d3d12_caps.SupportFlags & D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_ASYMETRIC_MOTION_PARTITION_SUPPORT) != 0)
|
||
codecSupport.hevc_support.hevc_features.bits.amp = PIPE_H265_ENC_FEATURE_SUPPORTED;
|
||
|
||
if ((codecSupport.hevc_support.d3d12_caps.SupportFlags & D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_ASYMETRIC_MOTION_PARTITION_REQUIRED) != 0)
|
||
codecSupport.hevc_support.hevc_features.bits.amp = (PIPE_H265_ENC_FEATURE_SUPPORTED | PIPE_H265_ENC_FEATURE_REQUIRED);
|
||
|
||
if ((codecSupport.hevc_support.d3d12_caps.SupportFlags & D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_SAO_FILTER_SUPPORT) != 0)
|
||
codecSupport.hevc_support.hevc_features.bits.sao = PIPE_H265_ENC_FEATURE_SUPPORTED;
|
||
|
||
if ((codecSupport.hevc_support.d3d12_caps.SupportFlags & D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_CONSTRAINED_INTRAPREDICTION_SUPPORT) != 0)
|
||
codecSupport.hevc_support.hevc_features.bits.constrained_intra_pred = PIPE_H265_ENC_FEATURE_SUPPORTED;
|
||
if ((codecSupport.hevc_support.d3d12_caps.SupportFlags & D3D12_VIDEO_ENCODER_CODEC_CONFIGURATION_SUPPORT_HEVC_FLAG_TRANSFORM_SKIP_SUPPORT) != 0)
|
||
codecSupport.hevc_support.hevc_features.bits.transform_skip = PIPE_H265_ENC_FEATURE_SUPPORTED;
|
||
|
||
DXGI_FORMAT encodeFormat = d3d12_convert_pipe_video_profile_to_dxgi_format(profile);
|
||
supportsProfile = supportsProfile &&
|
||
d3d12_video_encode_max_supported_resolution(codecDesc, maxRes, spD3D12VideoDevice.Get());
|
||
|
||
if (supportedSliceStructures == PIPE_VIDEO_CAP_SLICE_STRUCTURE_NONE)
|
||
maxSlices = 0;
|
||
else
|
||
supportsProfile = supportsProfile && d3d12_video_encode_max_supported_slices(codecDesc,
|
||
maxRes,
|
||
encodeFormat,
|
||
maxSlices,
|
||
spD3D12VideoDevice.Get(),
|
||
d3d12_codec_support);
|
||
}
|
||
} break;
|
||
default:
|
||
supportsProfile = false;
|
||
}
|
||
|
||
return VideoFeatureAreaSupport.VideoEncodeSupport && supportsProfile;
|
||
}
|
||
|
||
static int
|
||
d3d12_screen_get_video_param_decode(struct pipe_screen *pscreen,
|
||
enum pipe_video_profile profile,
|
||
enum pipe_video_entrypoint entrypoint,
|
||
enum pipe_video_cap param)
|
||
{
|
||
switch (param) {
|
||
case PIPE_VIDEO_CAP_REQUIRES_FLUSH_ON_END_FRAME:
|
||
/* As sometimes we need to copy the output
|
||
and sync with the context, we handle the
|
||
flush internally on end frame for decode
|
||
*/
|
||
return 0;
|
||
case PIPE_VIDEO_CAP_NPOT_TEXTURES:
|
||
return 1;
|
||
case PIPE_VIDEO_CAP_MAX_WIDTH:
|
||
case PIPE_VIDEO_CAP_MAX_HEIGHT:
|
||
case PIPE_VIDEO_CAP_MAX_LEVEL:
|
||
case PIPE_VIDEO_CAP_SUPPORTED:
|
||
{
|
||
if (d3d12_has_video_decode_support(pscreen, profile)) {
|
||
DXGI_FORMAT format = d3d12_convert_pipe_video_profile_to_dxgi_format(profile);
|
||
auto pipeFmt = d3d12_get_pipe_format(format);
|
||
bool formatSupported = pscreen->is_video_format_supported(pscreen, pipeFmt, profile, entrypoint);
|
||
if (formatSupported) {
|
||
GUID decodeGUID = d3d12_video_decoder_convert_pipe_video_profile_to_d3d12_profile(profile);
|
||
GUID emptyGUID = {};
|
||
if (decodeGUID != emptyGUID) {
|
||
bool supportAny = false;
|
||
D3D12_FEATURE_DATA_VIDEO_DECODE_SUPPORT outSupportedConfig = {};
|
||
D3D12_VIDEO_DECODE_CONFIGURATION decoderConfig = { decodeGUID,
|
||
D3D12_BITSTREAM_ENCRYPTION_TYPE_NONE,
|
||
D3D12_VIDEO_FRAME_CODED_INTERLACE_TYPE_NONE };
|
||
|
||
d3d12_video_resolution_to_level_mapping_entry bestSupportedConfig =
|
||
get_max_level_resolution_video_decode_support(decoderConfig,
|
||
format,
|
||
pscreen,
|
||
supportAny,
|
||
outSupportedConfig);
|
||
if (supportAny) {
|
||
if (param == PIPE_VIDEO_CAP_MAX_WIDTH) {
|
||
return bestSupportedConfig.resolution.Width;
|
||
} else if (param == PIPE_VIDEO_CAP_MAX_HEIGHT) {
|
||
return bestSupportedConfig.resolution.Height;
|
||
} else if (param == PIPE_VIDEO_CAP_MAX_LEVEL) {
|
||
return bestSupportedConfig.level;
|
||
} else if (param == PIPE_VIDEO_CAP_SUPPORTED) {
|
||
return 1;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
}
|
||
return 0;
|
||
} break;
|
||
case PIPE_VIDEO_CAP_PREFERED_FORMAT:
|
||
return (profile == PIPE_VIDEO_PROFILE_UNKNOWN) ? PIPE_FORMAT_NV12 : d3d12_get_pipe_format(d3d12_convert_pipe_video_profile_to_dxgi_format(profile));
|
||
case PIPE_VIDEO_CAP_PREFERS_INTERLACED:
|
||
return false;
|
||
case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
|
||
return true;
|
||
case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
|
||
return true;
|
||
case PIPE_VIDEO_CAP_SUPPORTS_CONTIGUOUS_PLANES_MAP:
|
||
return true;
|
||
break;
|
||
default:
|
||
debug_printf("[d3d12_screen_get_video_param] unknown video param: %d\n", param);
|
||
return 0;
|
||
}
|
||
}
|
||
|
||
|
||
static bool
|
||
d3d12_has_video_process_support(struct pipe_screen *pscreen, D3D12_FEATURE_DATA_VIDEO_PROCESS_SUPPORT &supportCaps)
|
||
{
|
||
ComPtr<ID3D12VideoDevice2> spD3D12VideoDevice;
|
||
struct d3d12_screen *pD3D12Screen = (struct d3d12_screen *) pscreen;
|
||
if (FAILED(pD3D12Screen->dev->QueryInterface(IID_PPV_ARGS(spD3D12VideoDevice.GetAddressOf())))) {
|
||
// No video encode support in underlying d3d12 device (needs ID3D12VideoDevice2)
|
||
return false;
|
||
}
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_FEATURE_AREA_SUPPORT VideoFeatureAreaSupport = {};
|
||
if (FAILED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_FEATURE_AREA_SUPPORT,
|
||
&VideoFeatureAreaSupport,
|
||
sizeof(VideoFeatureAreaSupport)))) {
|
||
return false;
|
||
}
|
||
|
||
struct ResolStruct {
|
||
uint Width;
|
||
uint Height;
|
||
};
|
||
|
||
ResolStruct resolutionsList[] = {
|
||
{ 8192, 8192 }, // 8k
|
||
{ 8192, 4320 }, // 8k - alternative
|
||
{ 7680, 4800 }, // 8k - alternative
|
||
{ 7680, 4320 }, // 8k - alternative
|
||
{ 4096, 2304 }, // 2160p (4K)
|
||
{ 4096, 2160 }, // 2160p (4K) - alternative
|
||
{ 2560, 1440 }, // 1440p
|
||
{ 1920, 1200 }, // 1200p
|
||
{ 1920, 1080 }, // 1080p
|
||
{ 1280, 720 }, // 720p
|
||
{ 800, 600 },
|
||
};
|
||
|
||
uint32_t idxResol = 0;
|
||
bool bSupportsAny = false;
|
||
while ((idxResol < ARRAY_SIZE(resolutionsList)) && !bSupportsAny) {
|
||
supportCaps.InputSample.Width = resolutionsList[idxResol].Width;
|
||
supportCaps.InputSample.Height = resolutionsList[idxResol].Height;
|
||
if (SUCCEEDED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_PROCESS_SUPPORT, &supportCaps, sizeof(supportCaps)))) {
|
||
bSupportsAny = ((supportCaps.SupportFlags & D3D12_VIDEO_PROCESS_SUPPORT_FLAG_SUPPORTED) != 0) ;
|
||
}
|
||
idxResol++;
|
||
}
|
||
|
||
return VideoFeatureAreaSupport.VideoProcessSupport && bSupportsAny;
|
||
}
|
||
|
||
static int
|
||
d3d12_screen_get_video_param_postproc(struct pipe_screen *pscreen,
|
||
enum pipe_video_profile profile,
|
||
enum pipe_video_entrypoint entrypoint,
|
||
enum pipe_video_cap param)
|
||
{
|
||
switch (param) {
|
||
case PIPE_VIDEO_CAP_REQUIRES_FLUSH_ON_END_FRAME:
|
||
return 1;
|
||
case PIPE_VIDEO_CAP_NPOT_TEXTURES:
|
||
return 1;
|
||
case PIPE_VIDEO_CAP_MAX_WIDTH:
|
||
case PIPE_VIDEO_CAP_MAX_HEIGHT:
|
||
case PIPE_VIDEO_CAP_SUPPORTED:
|
||
case PIPE_VIDEO_CAP_PREFERED_FORMAT:
|
||
case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
|
||
case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
|
||
case PIPE_VIDEO_CAP_SUPPORTS_CONTIGUOUS_PLANES_MAP:
|
||
case PIPE_VIDEO_CAP_VPP_MAX_INPUT_WIDTH:
|
||
case PIPE_VIDEO_CAP_VPP_MAX_INPUT_HEIGHT:
|
||
case PIPE_VIDEO_CAP_VPP_MIN_INPUT_WIDTH:
|
||
case PIPE_VIDEO_CAP_VPP_MIN_INPUT_HEIGHT:
|
||
case PIPE_VIDEO_CAP_VPP_MAX_OUTPUT_WIDTH:
|
||
case PIPE_VIDEO_CAP_VPP_MAX_OUTPUT_HEIGHT:
|
||
case PIPE_VIDEO_CAP_VPP_MIN_OUTPUT_WIDTH:
|
||
case PIPE_VIDEO_CAP_VPP_MIN_OUTPUT_HEIGHT:
|
||
case PIPE_VIDEO_CAP_VPP_ORIENTATION_MODES:
|
||
case PIPE_VIDEO_CAP_VPP_BLEND_MODES:
|
||
{
|
||
// Assume defaults for now, we don't have the input args passed by get_video_param to be accurate here.
|
||
const D3D12_VIDEO_FIELD_TYPE FieldType = D3D12_VIDEO_FIELD_TYPE_NONE;
|
||
const D3D12_VIDEO_FRAME_STEREO_FORMAT StereoFormat = D3D12_VIDEO_FRAME_STEREO_FORMAT_NONE;
|
||
const DXGI_RATIONAL FrameRate = { 30, 1 };
|
||
const DXGI_FORMAT InputFormat = DXGI_FORMAT_NV12;
|
||
const DXGI_COLOR_SPACE_TYPE InputColorSpace = DXGI_COLOR_SPACE_YCBCR_STUDIO_G22_LEFT_P709;
|
||
const DXGI_FORMAT OutputFormat = DXGI_FORMAT_NV12;
|
||
const DXGI_COLOR_SPACE_TYPE OutputColorSpace = DXGI_COLOR_SPACE_YCBCR_STUDIO_G22_LEFT_P709;
|
||
const UINT Width = 1280;
|
||
const UINT Height = 720;
|
||
D3D12_FEATURE_DATA_VIDEO_PROCESS_SUPPORT supportCaps =
|
||
{
|
||
0, // NodeIndex
|
||
{ Width, Height, { InputFormat, InputColorSpace } },
|
||
FieldType,
|
||
StereoFormat,
|
||
FrameRate,
|
||
{ OutputFormat, OutputColorSpace },
|
||
StereoFormat,
|
||
FrameRate,
|
||
};
|
||
|
||
if (d3d12_has_video_process_support(pscreen, supportCaps)) {
|
||
if (param == PIPE_VIDEO_CAP_SUPPORTED) {
|
||
return true;
|
||
} else if (param == PIPE_VIDEO_CAP_PREFERED_FORMAT) {
|
||
return PIPE_FORMAT_NV12;
|
||
} else if (param == PIPE_VIDEO_CAP_SUPPORTS_INTERLACED) {
|
||
return false;
|
||
} else if (param == PIPE_VIDEO_CAP_MAX_WIDTH) {
|
||
return supportCaps.InputSample.Width;
|
||
} else if (param == PIPE_VIDEO_CAP_MAX_HEIGHT) {
|
||
return supportCaps.InputSample.Height;
|
||
} else if (param == PIPE_VIDEO_CAP_SUPPORTS_CONTIGUOUS_PLANES_MAP) {
|
||
return true;
|
||
} else if (param == PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE) {
|
||
return true;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MAX_INPUT_WIDTH) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MaxWidth;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MAX_INPUT_HEIGHT) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MaxHeight;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MIN_INPUT_WIDTH) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MinWidth;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MIN_INPUT_HEIGHT) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MinHeight;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MAX_OUTPUT_WIDTH) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MaxWidth;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MAX_OUTPUT_HEIGHT) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MaxHeight;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MIN_OUTPUT_WIDTH) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MinWidth;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_MIN_OUTPUT_HEIGHT) {
|
||
return supportCaps.ScaleSupport.OutputSizeRange.MinHeight;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_BLEND_MODES) {
|
||
uint32_t blend_modes = PIPE_VIDEO_VPP_BLEND_MODE_NONE;
|
||
if (((supportCaps.FeatureSupport & D3D12_VIDEO_PROCESS_FEATURE_FLAG_ALPHA_BLENDING) != 0)
|
||
&& ((supportCaps.FeatureSupport & D3D12_VIDEO_PROCESS_FEATURE_FLAG_ALPHA_FILL) != 0))
|
||
{
|
||
blend_modes |= PIPE_VIDEO_VPP_BLEND_MODE_GLOBAL_ALPHA;
|
||
}
|
||
return blend_modes;
|
||
} else if (param == PIPE_VIDEO_CAP_VPP_ORIENTATION_MODES) {
|
||
uint32_t orientation_modes = PIPE_VIDEO_VPP_ORIENTATION_DEFAULT;
|
||
if((supportCaps.FeatureSupport & D3D12_VIDEO_PROCESS_FEATURE_FLAG_FLIP) != 0) {
|
||
orientation_modes |= PIPE_VIDEO_VPP_FLIP_HORIZONTAL;
|
||
orientation_modes |= PIPE_VIDEO_VPP_FLIP_VERTICAL;
|
||
}
|
||
|
||
if((supportCaps.FeatureSupport & D3D12_VIDEO_PROCESS_FEATURE_FLAG_ROTATION) != 0) {
|
||
orientation_modes |= PIPE_VIDEO_VPP_ROTATION_90;
|
||
orientation_modes |= PIPE_VIDEO_VPP_ROTATION_180;
|
||
orientation_modes |= PIPE_VIDEO_VPP_ROTATION_270;
|
||
}
|
||
return orientation_modes;
|
||
}
|
||
}
|
||
return 0;
|
||
} break;
|
||
default:
|
||
return 0;
|
||
}
|
||
}
|
||
|
||
static int
|
||
d3d12_screen_get_video_param_encode(struct pipe_screen *pscreen,
|
||
enum pipe_video_profile profile,
|
||
enum pipe_video_entrypoint entrypoint,
|
||
enum pipe_video_cap param)
|
||
{
|
||
uint32_t maxLvlEncode = 0u;
|
||
D3D12_VIDEO_ENCODER_PICTURE_RESOLUTION_DESC maxResEncode = {};
|
||
uint32_t maxSlices = 0u;
|
||
uint32_t supportedSliceStructures = 0u;
|
||
uint32_t maxReferencesPerFrame = 0u;
|
||
struct d3d12_encode_codec_support codec_specific_support;
|
||
memset(&codec_specific_support, 0, sizeof(codec_specific_support));
|
||
switch (param) {
|
||
case PIPE_VIDEO_CAP_ENC_SUPPORTS_ASYNC_OPERATION:
|
||
return D3D12_VIDEO_ENC_ASYNC;
|
||
case PIPE_VIDEO_CAP_REQUIRES_FLUSH_ON_END_FRAME:
|
||
return 1;
|
||
case PIPE_VIDEO_CAP_NPOT_TEXTURES:
|
||
return 1;
|
||
case PIPE_VIDEO_CAP_MAX_WIDTH:
|
||
case PIPE_VIDEO_CAP_MAX_HEIGHT:
|
||
case PIPE_VIDEO_CAP_MAX_LEVEL:
|
||
case PIPE_VIDEO_CAP_SUPPORTED:
|
||
case PIPE_VIDEO_CAP_ENC_MAX_SLICES_PER_FRAME:
|
||
case PIPE_VIDEO_CAP_ENC_SLICES_STRUCTURE:
|
||
case PIPE_VIDEO_CAP_ENC_MAX_REFERENCES_PER_FRAME:
|
||
case PIPE_VIDEO_CAP_ENC_HEVC_FEATURE_FLAGS:
|
||
case PIPE_VIDEO_CAP_ENC_HEVC_BLOCK_SIZES:
|
||
case PIPE_VIDEO_CAP_ENC_HEVC_PREDICTION_DIRECTION:
|
||
{
|
||
if (d3d12_has_video_encode_support(pscreen,
|
||
profile,
|
||
maxLvlEncode,
|
||
maxResEncode,
|
||
maxSlices,
|
||
supportedSliceStructures,
|
||
maxReferencesPerFrame,
|
||
codec_specific_support)) {
|
||
|
||
DXGI_FORMAT format = d3d12_convert_pipe_video_profile_to_dxgi_format(profile);
|
||
auto pipeFmt = d3d12_get_pipe_format(format);
|
||
bool formatSupported = pscreen->is_video_format_supported(pscreen, pipeFmt, profile, entrypoint);
|
||
if (formatSupported) {
|
||
if (param == PIPE_VIDEO_CAP_MAX_WIDTH) {
|
||
return maxResEncode.Width;
|
||
} else if (param == PIPE_VIDEO_CAP_MAX_HEIGHT) {
|
||
return maxResEncode.Height;
|
||
} else if (param == PIPE_VIDEO_CAP_MAX_LEVEL) {
|
||
return maxLvlEncode;
|
||
} else if (param == PIPE_VIDEO_CAP_SUPPORTED) {
|
||
return 1;
|
||
} else if (param == PIPE_VIDEO_CAP_ENC_MAX_SLICES_PER_FRAME) {
|
||
return maxSlices;
|
||
} else if (param == PIPE_VIDEO_CAP_ENC_SLICES_STRUCTURE) {
|
||
return supportedSliceStructures;
|
||
} else if (param == PIPE_VIDEO_CAP_ENC_MAX_REFERENCES_PER_FRAME) {
|
||
return maxReferencesPerFrame;
|
||
} else if (param == PIPE_VIDEO_CAP_ENC_HEVC_FEATURE_FLAGS) {
|
||
/* get_video_param sets hevc_features.bits.config_supported = 1
|
||
to distinguish between supported cap with all bits off and unsupported by driver
|
||
with value = 0
|
||
*/
|
||
return codec_specific_support.hevc_support.hevc_features.value;
|
||
} else if (param == PIPE_VIDEO_CAP_ENC_HEVC_BLOCK_SIZES) {
|
||
/* get_video_param sets hevc_block_sizes.bits.config_supported = 1
|
||
to distinguish between supported cap with all bits off and unsupported by driver
|
||
with value = 0
|
||
*/
|
||
return codec_specific_support.hevc_support.hevc_block_sizes.value;
|
||
} else if (param == PIPE_VIDEO_CAP_ENC_HEVC_PREDICTION_DIRECTION) {
|
||
return codec_specific_support.hevc_support.prediction_direction;
|
||
}
|
||
}
|
||
}
|
||
return 0;
|
||
} break;
|
||
case PIPE_VIDEO_CAP_PREFERED_FORMAT:
|
||
return (profile == PIPE_VIDEO_PROFILE_UNKNOWN) ? PIPE_FORMAT_NV12 : d3d12_get_pipe_format(d3d12_convert_pipe_video_profile_to_dxgi_format(profile));
|
||
case PIPE_VIDEO_CAP_PREFERS_INTERLACED:
|
||
return false;
|
||
case PIPE_VIDEO_CAP_SUPPORTS_INTERLACED:
|
||
return false;
|
||
case PIPE_VIDEO_CAP_SUPPORTS_PROGRESSIVE:
|
||
return true;
|
||
case PIPE_VIDEO_CAP_SUPPORTS_CONTIGUOUS_PLANES_MAP:
|
||
return true;
|
||
default:
|
||
debug_printf("[d3d12_screen_get_video_param] unknown video param: %d\n", param);
|
||
return 0;
|
||
}
|
||
}
|
||
|
||
static int
|
||
d3d12_screen_get_video_param(struct pipe_screen *pscreen,
|
||
enum pipe_video_profile profile,
|
||
enum pipe_video_entrypoint entrypoint,
|
||
enum pipe_video_cap param)
|
||
{
|
||
if (entrypoint == PIPE_VIDEO_ENTRYPOINT_BITSTREAM) {
|
||
return d3d12_screen_get_video_param_decode(pscreen, profile, entrypoint, param);
|
||
} else if (entrypoint == PIPE_VIDEO_ENTRYPOINT_ENCODE) {
|
||
return d3d12_screen_get_video_param_encode(pscreen, profile, entrypoint, param);
|
||
} else if (entrypoint == PIPE_VIDEO_ENTRYPOINT_PROCESSING) {
|
||
return d3d12_screen_get_video_param_postproc(pscreen, profile, entrypoint, param);
|
||
}
|
||
return 0;
|
||
}
|
||
|
||
static bool
|
||
is_d3d12_video_encode_format_supported(struct pipe_screen *screen,
|
||
pipe_format format,
|
||
enum pipe_video_profile profile)
|
||
{
|
||
D3D12_VIDEO_ENCODER_PROFILE_H264 profH264 = {};
|
||
D3D12_VIDEO_ENCODER_PROFILE_HEVC profHEVC = {};
|
||
D3D12_FEATURE_DATA_VIDEO_ENCODER_INPUT_FORMAT capDataFmt = {};
|
||
capDataFmt.NodeIndex = 0;
|
||
capDataFmt.Codec = d3d12_video_encoder_convert_codec_to_d3d12_enc_codec(profile);
|
||
capDataFmt.Format = d3d12_get_format(format);
|
||
switch (u_reduce_video_profile(profile)) {
|
||
case PIPE_VIDEO_FORMAT_MPEG4_AVC:
|
||
{
|
||
profH264 = d3d12_video_encoder_convert_profile_to_d3d12_enc_profile_h264(profile);
|
||
capDataFmt.Profile.DataSize = sizeof(profH264);
|
||
capDataFmt.Profile.pH264Profile = &profH264;
|
||
} break;
|
||
case PIPE_VIDEO_FORMAT_HEVC:
|
||
{
|
||
profHEVC = d3d12_video_encoder_convert_profile_to_d3d12_enc_profile_hevc(profile);
|
||
capDataFmt.Profile.DataSize = sizeof(profHEVC);
|
||
capDataFmt.Profile.pHEVCProfile = &profHEVC;
|
||
} break;
|
||
default:
|
||
{
|
||
unreachable("Unsupported pipe_video_format");
|
||
} break;
|
||
}
|
||
ComPtr<ID3D12VideoDevice3> spD3D12VideoDevice;
|
||
struct d3d12_screen *pD3D12Screen = (struct d3d12_screen *) screen;
|
||
if (FAILED(pD3D12Screen->dev->QueryInterface(IID_PPV_ARGS(spD3D12VideoDevice.GetAddressOf())))) {
|
||
// No video encode support in underlying d3d12 device (needs ID3D12VideoDevice3)
|
||
return false;
|
||
}
|
||
HRESULT hr = spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_ENCODER_INPUT_FORMAT,
|
||
&capDataFmt,
|
||
sizeof(capDataFmt));
|
||
return SUCCEEDED(hr) && capDataFmt.IsSupported;
|
||
}
|
||
|
||
static bool
|
||
is_d3d12_video_decode_format_supported(struct pipe_screen *screen,
|
||
pipe_format format,
|
||
enum pipe_video_profile profile)
|
||
{
|
||
ComPtr<ID3D12VideoDevice3> spD3D12VideoDevice;
|
||
struct d3d12_screen *pD3D12Screen = (struct d3d12_screen *) screen;
|
||
if (FAILED(pD3D12Screen->dev->QueryInterface(IID_PPV_ARGS(spD3D12VideoDevice.GetAddressOf()))))
|
||
return false; // No video encode support in underlying d3d12 device (needs ID3D12VideoDevice3)
|
||
|
||
GUID decodeGUID = d3d12_video_decoder_convert_pipe_video_profile_to_d3d12_profile(profile);
|
||
GUID emptyGUID = {};
|
||
assert (decodeGUID != emptyGUID);
|
||
|
||
D3D12_VIDEO_DECODE_CONFIGURATION decoderConfig = { decodeGUID,
|
||
D3D12_BITSTREAM_ENCRYPTION_TYPE_NONE,
|
||
D3D12_VIDEO_FRAME_CODED_INTERLACE_TYPE_NONE };
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_DECODE_FORMAT_COUNT decodeFormatCount = {0 /* NodeIndex*/, decoderConfig };
|
||
if(FAILED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_DECODE_FORMAT_COUNT,
|
||
&decodeFormatCount,
|
||
sizeof(decodeFormatCount))))
|
||
return false;
|
||
|
||
std::vector<DXGI_FORMAT> supportedDecodeFormats;
|
||
supportedDecodeFormats.resize(decodeFormatCount.FormatCount);
|
||
|
||
D3D12_FEATURE_DATA_VIDEO_DECODE_FORMATS decodeFormats =
|
||
{
|
||
0, // NodeIndex
|
||
decoderConfig,
|
||
static_cast<UINT>(supportedDecodeFormats.size()),
|
||
supportedDecodeFormats.data()
|
||
};
|
||
|
||
if(FAILED(spD3D12VideoDevice->CheckFeatureSupport(D3D12_FEATURE_VIDEO_DECODE_FORMATS,
|
||
&decodeFormats,
|
||
sizeof(decodeFormats))))
|
||
return false;
|
||
|
||
DXGI_FORMAT requestedDXGIFormat = d3d12_get_format(format);
|
||
for (DXGI_FORMAT fmt : supportedDecodeFormats)
|
||
if (fmt == requestedDXGIFormat)
|
||
return true;
|
||
return false;
|
||
}
|
||
|
||
static bool
|
||
is_d3d12_video_process_format_supported(struct pipe_screen *screen,
|
||
pipe_format format)
|
||
{
|
||
// Return both VPBlit support and format is in known list
|
||
return (screen->get_video_param(screen,
|
||
PIPE_VIDEO_PROFILE_UNKNOWN,
|
||
PIPE_VIDEO_ENTRYPOINT_PROCESSING,
|
||
PIPE_VIDEO_CAP_SUPPORTED))
|
||
&&
|
||
((format == PIPE_FORMAT_NV12) || (format == PIPE_FORMAT_P010)
|
||
|| (format == PIPE_FORMAT_R8G8B8A8_UNORM) || (format == PIPE_FORMAT_R8G8B8A8_UINT)
|
||
|| (format == PIPE_FORMAT_R8G8B8X8_UNORM) || (format == PIPE_FORMAT_R8G8B8X8_UINT));
|
||
}
|
||
|
||
static bool
|
||
is_d3d12_video_allowed_format(enum pipe_format format, enum pipe_video_entrypoint entrypoint)
|
||
{
|
||
if (entrypoint == PIPE_VIDEO_ENTRYPOINT_BITSTREAM) {
|
||
return ((format == PIPE_FORMAT_NV12) || (format == PIPE_FORMAT_P010));
|
||
} else if (entrypoint == PIPE_VIDEO_ENTRYPOINT_ENCODE) {
|
||
return ((format == PIPE_FORMAT_NV12) || (format == PIPE_FORMAT_P010));
|
||
} else if (entrypoint == PIPE_VIDEO_ENTRYPOINT_PROCESSING) {
|
||
return (format == PIPE_FORMAT_NV12) || (format == PIPE_FORMAT_P010)
|
||
|| (format == PIPE_FORMAT_R8G8B8A8_UNORM) || (format == PIPE_FORMAT_R8G8B8A8_UINT)
|
||
|| (format == PIPE_FORMAT_R8G8B8X8_UNORM) || (format == PIPE_FORMAT_R8G8B8X8_UINT);
|
||
}
|
||
return false;
|
||
}
|
||
|
||
static bool
|
||
d3d12_video_buffer_is_format_supported(struct pipe_screen *screen,
|
||
enum pipe_format format,
|
||
enum pipe_video_profile profile,
|
||
enum pipe_video_entrypoint entrypoint)
|
||
{
|
||
// Check in allowed list of formats first
|
||
if(!is_d3d12_video_allowed_format(format, entrypoint))
|
||
return false;
|
||
|
||
// If the VA frontend asks for all profiles, assign
|
||
// a default profile based on the bitdepth
|
||
if(u_reduce_video_profile(profile) == PIPE_VIDEO_FORMAT_UNKNOWN)
|
||
{
|
||
profile = (format == PIPE_FORMAT_P010) ? PIPE_VIDEO_PROFILE_HEVC_MAIN_10 : PIPE_VIDEO_PROFILE_MPEG4_AVC_MAIN;
|
||
}
|
||
|
||
// Then check is the underlying driver supports the allowed formats
|
||
if (entrypoint == PIPE_VIDEO_ENTRYPOINT_BITSTREAM) {
|
||
return is_d3d12_video_decode_format_supported(screen, format, profile);
|
||
} else if (entrypoint == PIPE_VIDEO_ENTRYPOINT_ENCODE) {
|
||
return is_d3d12_video_encode_format_supported(screen, format, profile);
|
||
} else if (entrypoint == PIPE_VIDEO_ENTRYPOINT_PROCESSING) {
|
||
return is_d3d12_video_process_format_supported(screen, format);
|
||
}
|
||
return false;
|
||
}
|
||
|
||
void
|
||
d3d12_screen_video_init(struct pipe_screen *pscreen)
|
||
{
|
||
pscreen->get_video_param = d3d12_screen_get_video_param;
|
||
pscreen->is_video_format_supported = d3d12_video_buffer_is_format_supported;
|
||
}
|