/* * Copyright 2022 Google LLC * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "include/gpu/graphite/vk/VulkanGraphiteUtils.h" #include "src/gpu/graphite/vk/VulkanGraphiteUtilsPriv.h" #include "include/gpu/ShaderErrorHandler.h" #include "include/gpu/graphite/Context.h" #include "include/gpu/vk/VulkanBackendContext.h" #include "src/core/SkTraceEvent.h" #include "src/gpu/graphite/ContextPriv.h" #include "src/gpu/graphite/vk/VulkanQueueManager.h" #include "src/gpu/graphite/vk/VulkanSampler.h" #include "src/gpu/graphite/vk/VulkanSharedContext.h" #include "src/sksl/SkSLProgramSettings.h" namespace skgpu::graphite::ContextFactory { std::unique_ptr MakeVulkan(const VulkanBackendContext& backendContext, const ContextOptions& options) { sk_sp sharedContext = VulkanSharedContext::Make(backendContext, options); if (!sharedContext) { return nullptr; } std::unique_ptr queueManager(new VulkanQueueManager(backendContext.fQueue, sharedContext.get())); if (!queueManager) { return nullptr; } return ContextCtorAccessor::MakeContext(std::move(sharedContext), std::move(queueManager), options); } } // namespace skgpu::graphite::ContextFactory namespace skgpu::graphite { VkShaderModule createVulkanShaderModule(const VulkanSharedContext* context, const std::string& spirv, VkShaderStageFlagBits stage) { TRACE_EVENT0("skia.shaders", "InstallVkShaderModule"); VkShaderModuleCreateInfo moduleCreateInfo; memset(&moduleCreateInfo, 0, sizeof(VkShaderModuleCreateInfo)); moduleCreateInfo.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO; moduleCreateInfo.pNext = nullptr; moduleCreateInfo.flags = 0; moduleCreateInfo.codeSize = spirv.size(); moduleCreateInfo.pCode = (const uint32_t*)spirv.c_str(); VkShaderModule shaderModule; VkResult result; VULKAN_CALL_RESULT(context, result, CreateShaderModule(context->device(), &moduleCreateInfo, /*const VkAllocationCallbacks*=*/nullptr, &shaderModule)); if (result != VK_SUCCESS) { SKGPU_LOG_E("Failed to create VkShaderModule"); return VK_NULL_HANDLE; } return shaderModule; } void DescriptorDataToVkDescSetLayout(const VulkanSharedContext* ctxt, const SkSpan& requestedDescriptors, VkDescriptorSetLayout* outLayout) { skia_private::STArray bindingLayouts; for (size_t i = 0; i < requestedDescriptors.size(); i++) { if (requestedDescriptors[i].fCount != 0) { const DescriptorData& currDescriptor = requestedDescriptors[i]; VkDescriptorSetLayoutBinding& layoutBinding = bindingLayouts.push_back(); memset(&layoutBinding, 0, sizeof(VkDescriptorSetLayoutBinding)); layoutBinding.binding = currDescriptor.fBindingIndex; layoutBinding.descriptorType = DsTypeEnumToVkDs(currDescriptor.fType); layoutBinding.descriptorCount = currDescriptor.fCount; layoutBinding.stageFlags = PipelineStageFlagsToVkShaderStageFlags(currDescriptor.fPipelineStageFlags); layoutBinding.pImmutableSamplers = currDescriptor.fImmutableSampler ? (static_cast( currDescriptor.fImmutableSampler))->constVkSamplerPtr() : nullptr; } } VkDescriptorSetLayoutCreateInfo layoutCreateInfo; memset(&layoutCreateInfo, 0, sizeof(VkDescriptorSetLayoutCreateInfo)); layoutCreateInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO; layoutCreateInfo.pNext = nullptr; layoutCreateInfo.flags = 0; layoutCreateInfo.bindingCount = bindingLayouts.size(); layoutCreateInfo.pBindings = &bindingLayouts.front(); VkResult result; VULKAN_CALL_RESULT( ctxt, result, CreateDescriptorSetLayout(ctxt->device(), &layoutCreateInfo, nullptr, outLayout)); if (result != VK_SUCCESS) { SkDebugf("Failed to create VkDescriptorSetLayout\n"); outLayout = VK_NULL_HANDLE; } } VkDescriptorType DsTypeEnumToVkDs(DescriptorType type) { switch (type) { case DescriptorType::kUniformBuffer: return VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC; case DescriptorType::kTextureSampler: return VK_DESCRIPTOR_TYPE_SAMPLER; case DescriptorType::kTexture: return VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE; case DescriptorType::kCombinedTextureSampler: return VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER; case DescriptorType::kStorageBuffer: return VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC; case DescriptorType::kInputAttachment: return VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT; } SkUNREACHABLE; } bool vkFormatIsSupported(VkFormat format) { switch (format) { case VK_FORMAT_R8G8B8A8_UNORM: case VK_FORMAT_B8G8R8A8_UNORM: case VK_FORMAT_R8G8B8A8_SRGB: case VK_FORMAT_R8G8B8_UNORM: case VK_FORMAT_R8G8_UNORM: case VK_FORMAT_A2B10G10R10_UNORM_PACK32: case VK_FORMAT_A2R10G10B10_UNORM_PACK32: case VK_FORMAT_R5G6B5_UNORM_PACK16: case VK_FORMAT_B4G4R4A4_UNORM_PACK16: case VK_FORMAT_R4G4B4A4_UNORM_PACK16: case VK_FORMAT_R8_UNORM: case VK_FORMAT_ETC2_R8G8B8_UNORM_BLOCK: case VK_FORMAT_BC1_RGB_UNORM_BLOCK: case VK_FORMAT_BC1_RGBA_UNORM_BLOCK: case VK_FORMAT_R16G16B16A16_SFLOAT: case VK_FORMAT_R16_SFLOAT: case VK_FORMAT_R16_UNORM: case VK_FORMAT_R16G16_UNORM: case VK_FORMAT_G8_B8_R8_3PLANE_420_UNORM: case VK_FORMAT_G8_B8R8_2PLANE_420_UNORM: case VK_FORMAT_R16G16B16A16_UNORM: case VK_FORMAT_R16G16_SFLOAT: case VK_FORMAT_S8_UINT: case VK_FORMAT_D16_UNORM: case VK_FORMAT_D32_SFLOAT: case VK_FORMAT_D24_UNORM_S8_UINT: case VK_FORMAT_D32_SFLOAT_S8_UINT: return true; default: return false; } } VkShaderStageFlags PipelineStageFlagsToVkShaderStageFlags( SkEnumBitMask stageFlags) { VkShaderStageFlags vkStageFlags = 0; if (stageFlags & PipelineStageFlags::kVertexShader) { vkStageFlags |= VK_SHADER_STAGE_VERTEX_BIT; } if (stageFlags & PipelineStageFlags::kFragmentShader) { vkStageFlags |= VK_SHADER_STAGE_FRAGMENT_BIT; } if (stageFlags & PipelineStageFlags::kCompute) { vkStageFlags |= VK_SHADER_STAGE_COMPUTE_BIT; } return vkStageFlags; } namespace ycbcrPackaging { uint32_t nonFormatInfoAsUInt32(const VulkanYcbcrConversionInfo& conversionInfo) { static_assert(kComponentAShift + kComponentBits <= 32); SkASSERT(conversionInfo.fYcbcrModel < (1u << kYcbcrModelBits )); SkASSERT(conversionInfo.fYcbcrRange < (1u << kYcbcrRangeBits )); SkASSERT(conversionInfo.fXChromaOffset < (1u << kXChromaOffsetBits )); SkASSERT(conversionInfo.fYChromaOffset < (1u << kYChromaOffsetBits )); SkASSERT(conversionInfo.fChromaFilter < (1u << kChromaFilterBits )); SkASSERT(conversionInfo.fForceExplicitReconstruction < (1u << kForceExplicitReconBits)); SkASSERT(conversionInfo.fComponents.r < (1u << kComponentBits )); SkASSERT(conversionInfo.fComponents.g < (1u << kComponentBits )); SkASSERT(conversionInfo.fComponents.b < (1u << kComponentBits )); SkASSERT(conversionInfo.fComponents.a < (1u << kComponentBits )); bool usesExternalFormat = conversionInfo.fFormat == VK_FORMAT_UNDEFINED; return (((uint32_t)(usesExternalFormat ) << kUsesExternalFormatShift) | ((uint32_t)(conversionInfo.fYcbcrModel ) << kYcbcrModelShift ) | ((uint32_t)(conversionInfo.fYcbcrRange ) << kYcbcrRangeShift ) | ((uint32_t)(conversionInfo.fXChromaOffset ) << kXChromaOffsetShift ) | ((uint32_t)(conversionInfo.fYChromaOffset ) << kYChromaOffsetShift ) | ((uint32_t)(conversionInfo.fChromaFilter ) << kChromaFilterShift ) | ((uint32_t)(conversionInfo.fForceExplicitReconstruction) << kForceExplicitReconShift) | ((uint32_t)(conversionInfo.fComponents.r ) << kComponentRShift ) | ((uint32_t)(conversionInfo.fComponents.g ) << kComponentGShift ) | ((uint32_t)(conversionInfo.fComponents.b ) << kComponentBShift ) | ((uint32_t)(conversionInfo.fComponents.a ) << kComponentAShift )); } int numInt32sNeeded(const VulkanYcbcrConversionInfo& conversionInfo) { if (!conversionInfo.isValid()) { return 0; } return conversionInfo.fFormat == VK_FORMAT_UNDEFINED ? SamplerDesc::kInt32sNeededExternalFormat : SamplerDesc::kInt32sNeededKnownFormat; } } // namespace ycbcrPackaging } // namespace skgpu::graphite