blob: b6af74f3b0fba3eb56c7491c54ace2f80eeefd32 [file] [log] [blame]
/*
* Copyright 2022 Google LLC
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "src/gpu/graphite/vk/VulkanGraphiteUtils.h"
#include "include/core/SkStream.h"
#include "include/gpu/ShaderErrorHandler.h"
#include "include/gpu/graphite/Context.h"
#include "include/gpu/vk/VulkanBackendContext.h"
#include "src/core/SkTraceEvent.h"
#include "src/gpu/graphite/ContextPriv.h"
#include "src/gpu/graphite/RenderPassDesc.h"
#include "src/gpu/graphite/TextureFormat.h"
#include "src/gpu/graphite/vk/VulkanQueueManager.h"
#include "src/gpu/graphite/vk/VulkanSampler.h"
#include "src/gpu/graphite/vk/VulkanSharedContext.h"
#include "src/sksl/SkSLProgramSettings.h"
namespace skgpu::graphite::ContextFactory {
std::unique_ptr<Context> MakeVulkan(const VulkanBackendContext& backendContext,
const ContextOptions& options) {
sk_sp<SharedContext> sharedContext = VulkanSharedContext::Make(backendContext, options);
if (!sharedContext) {
return nullptr;
}
std::unique_ptr<QueueManager> queueManager(new VulkanQueueManager(backendContext.fQueue,
sharedContext.get()));
if (!queueManager) {
return nullptr;
}
return ContextCtorAccessor::MakeContext(std::move(sharedContext),
std::move(queueManager),
options);
}
} // namespace skgpu::graphite::ContextFactory
namespace skgpu::graphite {
VkShaderModule CreateVulkanShaderModule(const VulkanSharedContext* context,
const std::string& spirv,
VkShaderStageFlagBits stage) {
TRACE_EVENT0("skia.shaders", "InstallVkShaderModule");
VkShaderModuleCreateInfo moduleCreateInfo = {};
moduleCreateInfo.sType = VK_STRUCTURE_TYPE_SHADER_MODULE_CREATE_INFO;
moduleCreateInfo.codeSize = spirv.size();
moduleCreateInfo.pCode = (const uint32_t*)spirv.c_str();
VkShaderModule shaderModule;
VkResult result;
VULKAN_CALL_RESULT(context,
result,
CreateShaderModule(context->device(),
&moduleCreateInfo,
/*const VkAllocationCallbacks*=*/nullptr,
&shaderModule));
if (result != VK_SUCCESS) {
SKGPU_LOG_E("Failed to create VkShaderModule");
return VK_NULL_HANDLE;
}
return shaderModule;
}
void DescriptorDataToVkDescSetLayout(const VulkanSharedContext* ctxt,
const SkSpan<DescriptorData>& requestedDescriptors,
VkDescriptorSetLayout* outLayout) {
// If requestedDescriptors is empty, that simply means we should create an empty placeholder
// layout that doesn't actually contain any descriptors.
skia_private::STArray<kDescriptorTypeCount, VkDescriptorSetLayoutBinding> bindingLayouts;
for (size_t i = 0; i < requestedDescriptors.size(); i++) {
if (requestedDescriptors[i].fCount != 0) {
const DescriptorData& currDescriptor = requestedDescriptors[i];
VkDescriptorSetLayoutBinding& layoutBinding = bindingLayouts.push_back();
layoutBinding = {};
layoutBinding.binding = currDescriptor.fBindingIndex;
layoutBinding.descriptorType = DsTypeEnumToVkDs(currDescriptor.fType);
layoutBinding.descriptorCount = currDescriptor.fCount;
layoutBinding.stageFlags =
PipelineStageFlagsToVkShaderStageFlags(currDescriptor.fPipelineStageFlags);
layoutBinding.pImmutableSamplers = currDescriptor.fImmutableSampler
? (static_cast<const VulkanSampler*>(
currDescriptor.fImmutableSampler))->constVkSamplerPtr()
: nullptr;
}
}
VkDescriptorSetLayoutCreateInfo layoutCreateInfo = {};
layoutCreateInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO;
layoutCreateInfo.bindingCount = bindingLayouts.size();
layoutCreateInfo.pBindings = bindingLayouts.data();
VkResult result;
VULKAN_CALL_RESULT(
ctxt,
result,
CreateDescriptorSetLayout(ctxt->device(), &layoutCreateInfo, nullptr, outLayout));
if (result != VK_SUCCESS) {
SkDebugf("Failed to create VkDescriptorSetLayout\n");
outLayout = VK_NULL_HANDLE;
}
}
VkDescriptorType DsTypeEnumToVkDs(DescriptorType type) {
switch (type) {
case DescriptorType::kUniformBuffer:
return VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC;
case DescriptorType::kTextureSampler:
return VK_DESCRIPTOR_TYPE_SAMPLER;
case DescriptorType::kTexture:
return VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE;
case DescriptorType::kCombinedTextureSampler:
return VK_DESCRIPTOR_TYPE_COMBINED_IMAGE_SAMPLER;
case DescriptorType::kStorageBuffer:
return VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC;
case DescriptorType::kInputAttachment:
return VK_DESCRIPTOR_TYPE_INPUT_ATTACHMENT;
}
SkUNREACHABLE;
}
#define VK_FORMAT_MAPPING(M) \
M(TextureFormat::kR8, VK_FORMAT_R8_UNORM) \
M(TextureFormat::kR16, VK_FORMAT_R16_UNORM) \
M(TextureFormat::kR16F, VK_FORMAT_R16_SFLOAT) \
M(TextureFormat::kR32F, VK_FORMAT_R32_SFLOAT) \
/*TextureFormat::kA8 unsupported */ \
M(TextureFormat::kRG8, VK_FORMAT_R8G8_UNORM) \
M(TextureFormat::kRG16, VK_FORMAT_R16G16_UNORM) \
M(TextureFormat::kRG16F, VK_FORMAT_R16G16_SFLOAT) \
M(TextureFormat::kRG32F, VK_FORMAT_R32G32_SFLOAT) \
M(TextureFormat::kRGB8, VK_FORMAT_R8G8B8_UNORM) \
M(TextureFormat::kBGR8, VK_FORMAT_B8G8R8_UNORM) \
M(TextureFormat::kB5_G6_R5, VK_FORMAT_R5G6B5_UNORM_PACK16) \
M(TextureFormat::kR5_G6_B5, VK_FORMAT_B5G6R5_UNORM_PACK16) \
M(TextureFormat::kRGB16, VK_FORMAT_R16G16B16_UNORM) \
M(TextureFormat::kRGB16F, VK_FORMAT_R16G16B16_SFLOAT) \
M(TextureFormat::kRGB32F, VK_FORMAT_R32G32B32_SFLOAT) \
M(TextureFormat::kRGB8_sRGB, VK_FORMAT_R8G8B8_SRGB) \
/*TextureFormat::kBGR10_XR unsupported */ \
M(TextureFormat::kRGBA8, VK_FORMAT_R8G8B8A8_UNORM) \
M(TextureFormat::kRGBA16, VK_FORMAT_R16G16B16A16_UNORM) \
M(TextureFormat::kRGBA16F, VK_FORMAT_R16G16B16A16_SFLOAT) \
M(TextureFormat::kRGBA32F, VK_FORMAT_R32G32B32A32_SFLOAT) \
M(TextureFormat::kRGB10_A2, VK_FORMAT_A2B10G10R10_UNORM_PACK32) \
M(TextureFormat::kRGBA8_sRGB, VK_FORMAT_R8G8B8A8_SRGB) \
M(TextureFormat::kBGRA8, VK_FORMAT_B8G8R8A8_UNORM) \
M(TextureFormat::kBGR10_A2, VK_FORMAT_A2R10G10B10_UNORM_PACK32) \
M(TextureFormat::kBGRA8_sRGB, VK_FORMAT_B8G8R8A8_SRGB) \
M(TextureFormat::kABGR4, VK_FORMAT_R4G4B4A4_UNORM_PACK16) \
M(TextureFormat::kARGB4, VK_FORMAT_B4G4R4A4_UNORM_PACK16) \
/*TextureFormat::kBGRA10x6_XR unsupported */ \
M(TextureFormat::kRGB8_ETC2, VK_FORMAT_ETC2_R8G8B8_UNORM_BLOCK) \
M(TextureFormat::kRGB8_ETC2_sRGB, VK_FORMAT_ETC2_R8G8B8_SRGB_BLOCK) \
M(TextureFormat::kRGB8_BC1, VK_FORMAT_BC1_RGB_UNORM_BLOCK) \
M(TextureFormat::kRGBA8_BC1, VK_FORMAT_BC1_RGBA_UNORM_BLOCK) \
M(TextureFormat::kRGBA8_BC1_sRGB, VK_FORMAT_BC1_RGBA_SRGB_BLOCK) \
M(TextureFormat::kYUV8_P2_420, VK_FORMAT_G8_B8R8_2PLANE_420_UNORM) \
M(TextureFormat::kYUV8_P3_420, VK_FORMAT_G8_B8_R8_3PLANE_420_UNORM) \
M(TextureFormat::kYUV10x6_P2_420, VK_FORMAT_G10X6_B10X6R10X6_2PLANE_420_UNORM_3PACK16) \
/*TextureFormat::kExternal VK_FORMAT_UNDEFINED w/ uint64_t sidecar */ \
M(TextureFormat::kS8, VK_FORMAT_S8_UINT) \
M(TextureFormat::kD16, VK_FORMAT_D16_UNORM) \
M(TextureFormat::kD32F, VK_FORMAT_D32_SFLOAT) \
M(TextureFormat::kD24_S8, VK_FORMAT_D24_UNORM_S8_UINT) \
M(TextureFormat::kD32F_S8, VK_FORMAT_D32_SFLOAT_S8_UINT)
// NOTE: For external formats, Vulkan stores the value outside of VkFormat and expects the VkFormat
// to be VK_FORMAT_UNDEFINED. The mapping functions below are unaware of this side car, so it
// maps VK_FORMAT_UNDEFINED to TextureFormat::kUnsupported (this is disambiguated explicitly by
// VulkanTextureInfo which has the full context). The reverse maps TextureFormat::kExternal to
// VK_FORMAT_UNDEFINED, as expected, although callers are responsible for passing along the actual
// external format separately.
TextureFormat VkFormatToTextureFormat(VkFormat format) {
#define M(TF, VK) case VK: return TF;
switch(format) {
VK_FORMAT_MAPPING(M)
default: return TextureFormat::kUnsupported;
}
#undef M
}
VkFormat TextureFormatToVkFormat(TextureFormat format) {
#define M(TF, VK) case TF: return VK;
switch(format) {
VK_FORMAT_MAPPING(M)
default: return VK_FORMAT_UNDEFINED;
}
#undef M
}
VkShaderStageFlags PipelineStageFlagsToVkShaderStageFlags(
SkEnumBitMask<PipelineStageFlags> stageFlags) {
VkShaderStageFlags vkStageFlags = 0;
if (stageFlags & PipelineStageFlags::kVertexShader) {
vkStageFlags |= VK_SHADER_STAGE_VERTEX_BIT;
}
if (stageFlags & PipelineStageFlags::kFragmentShader) {
vkStageFlags |= VK_SHADER_STAGE_FRAGMENT_BIT;
}
if (stageFlags & PipelineStageFlags::kCompute) {
vkStageFlags |= VK_SHADER_STAGE_COMPUTE_BIT;
}
return vkStageFlags;
}
bool RenderPassDescWillLoadMSAAFromResolve(const RenderPassDesc& renderPassDesc) {
return renderPassDesc.fColorResolveAttachment.fFormat != TextureFormat::kUnsupported &&
renderPassDesc.fColorResolveAttachment.fLoadOp == LoadOp::kLoad;
}
} // namespace skgpu::graphite