| /* |
| * Copyright 2022 Google LLC |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| |
| #include "src/gpu/graphite/vk/VulkanTexture.h" |
| |
| #include "include/gpu/MutableTextureState.h" |
| #include "include/gpu/graphite/vk/VulkanGraphiteTypes.h" |
| #include "include/gpu/vk/VulkanMutableTextureState.h" |
| #include "src/core/SkCompressedDataUtils.h" |
| #include "src/core/SkMipmap.h" |
| #include "src/gpu/DataUtils.h" |
| #include "src/gpu/graphite/Log.h" |
| #include "src/gpu/graphite/Sampler.h" |
| #include "src/gpu/graphite/task/UploadTask.h" |
| #include "src/gpu/graphite/vk/VulkanCaps.h" |
| #include "src/gpu/graphite/vk/VulkanCommandBuffer.h" |
| #include "src/gpu/graphite/vk/VulkanDescriptorSet.h" |
| #include "src/gpu/graphite/vk/VulkanFramebuffer.h" |
| #include "src/gpu/graphite/vk/VulkanGraphiteUtils.h" |
| #include "src/gpu/graphite/vk/VulkanResourceProvider.h" |
| #include "src/gpu/graphite/vk/VulkanSharedContext.h" |
| #include "src/gpu/vk/VulkanMemory.h" |
| #include "src/gpu/vk/VulkanMutableTextureStatePriv.h" |
| |
| using namespace skia_private; |
| |
| namespace skgpu::graphite { |
| |
| bool VulkanTexture::MakeVkImage(const VulkanSharedContext* sharedContext, |
| SkISize dimensions, |
| const TextureInfo& info, |
| CreatedImageInfo* outInfo) { |
| SkASSERT(outInfo); |
| const VulkanCaps& caps = sharedContext->vulkanCaps(); |
| |
| if (dimensions.isEmpty()) { |
| SKGPU_LOG_E("Tried to create VkImage with empty dimensions."); |
| return false; |
| } |
| if (dimensions.width() > caps.maxTextureSize() || |
| dimensions.height() > caps.maxTextureSize()) { |
| SKGPU_LOG_E("Tried to create VkImage with too large a size."); |
| return false; |
| } |
| |
| if ((info.isProtected() == Protected::kYes) != caps.protectedSupport()) { |
| SKGPU_LOG_E("Tried to create %s VkImage in %s Context.", |
| info.isProtected() == Protected::kYes ? "protected" : "unprotected", |
| caps.protectedSupport() ? "protected" : "unprotected"); |
| return false; |
| } |
| |
| const auto& vkInfo = TextureInfoPriv::Get<VulkanTextureInfo>(info); |
| |
| bool isLinear = vkInfo.fImageTiling == VK_IMAGE_TILING_LINEAR; |
| VkImageLayout initialLayout = isLinear ? VK_IMAGE_LAYOUT_PREINITIALIZED |
| : VK_IMAGE_LAYOUT_UNDEFINED; |
| |
| // Create Image |
| VkSampleCountFlagBits vkSamples; |
| if (!SampleCountToVkSampleCount(vkInfo.fSampleCount, &vkSamples)) { |
| SKGPU_LOG_E("Failed creating VkImage because we could not covert the number of samples: " |
| "%u to a VkSampleCountFlagBits.", info.numSamples()); |
| return false; |
| } |
| |
| SkASSERT(!isLinear || vkSamples == VK_SAMPLE_COUNT_1_BIT); |
| SkASSERT(info.isProtected() == Protected::kNo || |
| (caps.protectedSupport() && SkToBool(VK_IMAGE_CREATE_PROTECTED_BIT & vkInfo.fFlags))); |
| |
| uint32_t numMipLevels = 1; |
| if (vkInfo.fMipmapped == Mipmapped::kYes) { |
| numMipLevels = SkMipmap::ComputeLevelCount(dimensions) + 1; |
| } |
| |
| uint32_t width = static_cast<uint32_t>(dimensions.fWidth); |
| uint32_t height = static_cast<uint32_t>(dimensions.fHeight); |
| |
| const VkImageCreateInfo imageCreateInfo = { |
| VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO, // sType |
| nullptr, // pNext |
| vkInfo.fFlags, // VkImageCreateFlags |
| VK_IMAGE_TYPE_2D, // VkImageType |
| vkInfo.fFormat, // VkFormat |
| { width, height, 1 }, // VkExtent3D |
| numMipLevels, // mipLevels |
| 1, // arrayLayers |
| vkSamples, // samples |
| vkInfo.fImageTiling, // VkImageTiling |
| vkInfo.fImageUsageFlags, // VkImageUsageFlags |
| vkInfo.fSharingMode, // VkSharingMode |
| 0, // queueFamilyCount |
| nullptr, // pQueueFamilyIndices |
| initialLayout // initialLayout |
| }; |
| |
| auto device = sharedContext->device(); |
| |
| VkImage image = VK_NULL_HANDLE; |
| VkResult result; |
| VULKAN_CALL_RESULT( |
| sharedContext, result, CreateImage(device, &imageCreateInfo, nullptr, &image)); |
| if (result != VK_SUCCESS) { |
| SKGPU_LOG_E("Failed call to vkCreateImage with error: %d", result); |
| return false; |
| } |
| |
| auto allocator = sharedContext->memoryAllocator(); |
| bool forceDedicatedMemory = caps.shouldAlwaysUseDedicatedImageMemory(); |
| bool useLazyAllocation = |
| SkToBool(vkInfo.fImageUsageFlags & VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT); |
| |
| auto checkResult = [sharedContext](VkResult result) { |
| return sharedContext->checkVkResult(result); |
| }; |
| if (!skgpu::VulkanMemory::AllocImageMemory(allocator, |
| image, |
| info.isProtected(), |
| forceDedicatedMemory, |
| useLazyAllocation, |
| checkResult, |
| &outInfo->fMemoryAlloc)) { |
| // If lazy memory allocation fails, fallback to attempting to use a regular allocation. |
| if (useLazyAllocation && |
| skgpu::VulkanMemory::AllocImageMemory(allocator, |
| image, |
| info.isProtected(), |
| forceDedicatedMemory, |
| /*useLazyAllocation=*/false, |
| checkResult, |
| &outInfo->fMemoryAlloc)) { |
| SKGPU_LOG_W("Could not allocate lazy image memory; using non-lazy instead."); |
| useLazyAllocation = false; |
| } else { |
| const char* protectednessStr = |
| info.isProtected() == Protected::kYes ? "protected" : "unprotected"; |
| const char* memoryTypeStr = forceDedicatedMemory ? "dedicated" : "shared"; |
| SKGPU_LOG_E("Failed to allocate %s %s image memory.", protectednessStr, memoryTypeStr); |
| |
| VULKAN_CALL(sharedContext->interface(), DestroyImage(device, image, nullptr)); |
| return false; |
| } |
| } |
| |
| VULKAN_CALL_RESULT( |
| sharedContext, |
| result, |
| BindImageMemory( |
| device, image, outInfo->fMemoryAlloc.fMemory, outInfo->fMemoryAlloc.fOffset)); |
| if (result != VK_SUCCESS) { |
| skgpu::VulkanMemory::FreeImageMemory(allocator, outInfo->fMemoryAlloc); |
| VULKAN_CALL(sharedContext->interface(), DestroyImage(device, image, nullptr)); |
| return false; |
| } |
| |
| outInfo->fImage = image; |
| outInfo->fMutableState = sk_make_sp<MutableTextureState>( |
| skgpu::MutableTextureStates::MakeVulkan(initialLayout, VK_QUEUE_FAMILY_IGNORED)); |
| return true; |
| } |
| |
| sk_sp<Texture> VulkanTexture::Make(const VulkanSharedContext* sharedContext, |
| SkISize dimensions, |
| const TextureInfo& info, |
| sk_sp<VulkanYcbcrConversion> ycbcrConversion) { |
| CreatedImageInfo imageInfo; |
| if (!MakeVkImage(sharedContext, dimensions, info, &imageInfo)) { |
| return nullptr; |
| } |
| |
| return sk_sp<Texture>(new VulkanTexture(sharedContext, |
| dimensions, |
| info, |
| std::move(imageInfo.fMutableState), |
| imageInfo.fImage, |
| imageInfo.fMemoryAlloc, |
| Ownership::kOwned, |
| std::move(ycbcrConversion))); |
| } |
| |
| sk_sp<Texture> VulkanTexture::MakeWrapped(const VulkanSharedContext* sharedContext, |
| SkISize dimensions, |
| const TextureInfo& info, |
| sk_sp<MutableTextureState> mutableState, |
| VkImage image, |
| const VulkanAlloc& alloc, |
| sk_sp<VulkanYcbcrConversion> ycbcrConversion) { |
| return sk_sp<Texture>(new VulkanTexture(sharedContext, |
| dimensions, |
| info, |
| std::move(mutableState), |
| image, |
| alloc, |
| Ownership::kWrapped, |
| std::move(ycbcrConversion))); |
| } |
| |
| VulkanTexture::~VulkanTexture() {} |
| |
| void VulkanTexture::setImageLayoutAndQueueIndex(VulkanCommandBuffer* cmdBuffer, |
| VkImageLayout newLayout, |
| VkAccessFlags dstAccessMask, |
| VkPipelineStageFlags dstStageMask, |
| uint32_t newQueueFamilyIndex) const { |
| |
| SkASSERT(newLayout == this->currentLayout() || |
| (VK_IMAGE_LAYOUT_UNDEFINED != newLayout && |
| VK_IMAGE_LAYOUT_PREINITIALIZED != newLayout)); |
| VkImageLayout currentLayout = this->currentLayout(); |
| uint32_t currentQueueIndex = this->currentQueueFamilyIndex(); |
| |
| const auto& textureInfo = this->vulkanTextureInfo(); |
| auto sharedContext = static_cast<const VulkanSharedContext*>(this->sharedContext()); |
| |
| // Enable the following block on new devices to test that their lazy images stay at 0 memory use |
| #if 0 |
| auto device = sharedContext->device(); |
| if (fAlloc.fFlags & skgpu::VulkanAlloc::kLazilyAllocated_Flag) { |
| VkDeviceSize size; |
| VULKAN_CALL(sharedContext->interface(), |
| GetDeviceMemoryCommitment(device, fAlloc.fMemory, &size)); |
| SkDebugf("Lazy Image. This: %p, image: %d, size: %d\n", this, fImage, size); |
| } |
| #endif |
| #ifdef SK_DEBUG |
| if (textureInfo.fSharingMode == VK_SHARING_MODE_CONCURRENT) { |
| if (newQueueFamilyIndex == VK_QUEUE_FAMILY_IGNORED) { |
| SkASSERT(currentQueueIndex == VK_QUEUE_FAMILY_IGNORED || |
| currentQueueIndex == VK_QUEUE_FAMILY_EXTERNAL || |
| currentQueueIndex == VK_QUEUE_FAMILY_FOREIGN_EXT); |
| } else { |
| SkASSERT(newQueueFamilyIndex == VK_QUEUE_FAMILY_EXTERNAL || |
| newQueueFamilyIndex == VK_QUEUE_FAMILY_FOREIGN_EXT); |
| SkASSERT(currentQueueIndex == VK_QUEUE_FAMILY_IGNORED); |
| } |
| } else { |
| SkASSERT(textureInfo.fSharingMode == VK_SHARING_MODE_EXCLUSIVE); |
| if (newQueueFamilyIndex == VK_QUEUE_FAMILY_IGNORED || |
| currentQueueIndex == sharedContext->queueIndex()) { |
| SkASSERT(currentQueueIndex == VK_QUEUE_FAMILY_IGNORED || |
| currentQueueIndex == VK_QUEUE_FAMILY_EXTERNAL || |
| currentQueueIndex == VK_QUEUE_FAMILY_FOREIGN_EXT || |
| currentQueueIndex == sharedContext->queueIndex()); |
| } else if (newQueueFamilyIndex == VK_QUEUE_FAMILY_EXTERNAL || |
| newQueueFamilyIndex == VK_QUEUE_FAMILY_FOREIGN_EXT) { |
| SkASSERT(currentQueueIndex == VK_QUEUE_FAMILY_IGNORED || |
| currentQueueIndex == sharedContext->queueIndex()); |
| } |
| } |
| #endif |
| |
| if (textureInfo.fSharingMode == VK_SHARING_MODE_EXCLUSIVE) { |
| if (newQueueFamilyIndex == VK_QUEUE_FAMILY_IGNORED) { |
| newQueueFamilyIndex = sharedContext->queueIndex(); |
| } |
| if (currentQueueIndex == VK_QUEUE_FAMILY_IGNORED) { |
| currentQueueIndex = sharedContext->queueIndex(); |
| } |
| } |
| |
| // If the old and new layout are the same and the layout is a read only layout, there is no need |
| // to put in a barrier unless we also need to switch queues. |
| if (newLayout == currentLayout && currentQueueIndex == newQueueFamilyIndex && |
| (VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL == currentLayout || |
| VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL == currentLayout || |
| VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL == currentLayout)) { |
| return; |
| } |
| |
| VkAccessFlags srcAccessMask = VulkanTexture::LayoutToSrcAccessMask(currentLayout); |
| VkPipelineStageFlags srcStageMask = VulkanTexture::LayoutToPipelineSrcStageFlags(currentLayout); |
| |
| VkImageAspectFlags aspectFlags = |
| GetVkImageAspectFlags(TextureInfoPriv::ViewFormat(this->textureInfo())); |
| uint32_t numMipLevels = 1; |
| SkISize dimensions = this->dimensions(); |
| if (this->mipmapped() == Mipmapped::kYes) { |
| numMipLevels = SkMipmap::ComputeLevelCount(dimensions) + 1; |
| } |
| VkImageMemoryBarrier imageMemoryBarrier = { |
| VK_STRUCTURE_TYPE_IMAGE_MEMORY_BARRIER, // sType |
| nullptr, // pNext |
| srcAccessMask, // srcAccessMask |
| dstAccessMask, // dstAccessMask |
| currentLayout, // oldLayout |
| newLayout, // newLayout |
| currentQueueIndex, // srcQueueFamilyIndex |
| newQueueFamilyIndex, // dstQueueFamilyIndex |
| fImage, // image |
| { aspectFlags, 0, numMipLevels, 0, 1 } // subresourceRange |
| }; |
| SkASSERT(srcAccessMask == imageMemoryBarrier.srcAccessMask); |
| cmdBuffer->addImageMemoryBarrier(this, srcStageMask, dstStageMask, /*byRegion=*/false, |
| &imageMemoryBarrier); |
| |
| skgpu::MutableTextureStates::SetVkImageLayout(this->mutableState(), newLayout); |
| skgpu::MutableTextureStates::SetVkQueueFamilyIndex(this->mutableState(), newQueueFamilyIndex); |
| } |
| |
| namespace { |
| |
| bool uses_lazy_memory(const VulkanAlloc& alloc) { |
| return alloc.fFlags & VulkanAlloc::Flag::kLazilyAllocated_Flag; |
| } |
| |
| #ifdef SK_DEBUG |
| bool has_transient_usage(const TextureInfo& info) { |
| const auto& vkInfo = TextureInfoPriv::Get<VulkanTextureInfo>(info); |
| return vkInfo.fImageUsageFlags & VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT; |
| } |
| #endif |
| |
| } // anonymous |
| |
| VulkanTexture::VulkanTexture(const VulkanSharedContext* sharedContext, |
| SkISize dimensions, |
| const TextureInfo& info, |
| sk_sp<MutableTextureState> mutableState, |
| VkImage image, |
| const VulkanAlloc& alloc, |
| Ownership ownership, |
| sk_sp<VulkanYcbcrConversion> ycbcrConversion) |
| : Texture(sharedContext, |
| dimensions, |
| info, |
| uses_lazy_memory(alloc), |
| std::move(mutableState), |
| ownership) |
| , fImage(image) |
| , fMemoryAlloc(alloc) |
| , fYcbcrConversion(std::move(ycbcrConversion)) { |
| SkASSERT(!uses_lazy_memory(fMemoryAlloc) || has_transient_usage(info)); |
| } |
| |
| void VulkanTexture::freeGpuData() { |
| // Need to delete any ImageViews first |
| fImageViews.clear(); |
| |
| // If the texture is wrapped we don't own this data |
| if (this->ownership() != Ownership::kWrapped) { |
| auto sharedContext = static_cast<const VulkanSharedContext*>(this->sharedContext()); |
| VULKAN_CALL(sharedContext->interface(), |
| DestroyImage(sharedContext->device(), fImage, nullptr)); |
| skgpu::VulkanMemory::FreeImageMemory(sharedContext->memoryAllocator(), fMemoryAlloc); |
| } |
| } |
| |
| void VulkanTexture::updateImageLayout(VkImageLayout newLayout) { |
| skgpu::MutableTextureStates::SetVkImageLayout(this->mutableState(), newLayout); |
| } |
| |
| VkImageLayout VulkanTexture::currentLayout() const { |
| return skgpu::MutableTextureStates::GetVkImageLayout(this->mutableState()); |
| } |
| |
| uint32_t VulkanTexture::currentQueueFamilyIndex() const { |
| return skgpu::MutableTextureStates::GetVkQueueFamilyIndex(this->mutableState()); |
| } |
| |
| VkPipelineStageFlags VulkanTexture::LayoutToPipelineSrcStageFlags(const VkImageLayout layout) { |
| if (VK_IMAGE_LAYOUT_GENERAL == layout) { |
| return VK_PIPELINE_STAGE_ALL_COMMANDS_BIT; |
| } else if (VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL == layout || |
| VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL == layout) { |
| return VK_PIPELINE_STAGE_TRANSFER_BIT; |
| } else if (VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL == layout) { |
| return VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT; |
| } else if (VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL == layout || |
| VK_IMAGE_LAYOUT_DEPTH_STENCIL_READ_ONLY_OPTIMAL == layout) { |
| return VK_PIPELINE_STAGE_LATE_FRAGMENT_TESTS_BIT; |
| } else if (VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL == layout) { |
| return VK_PIPELINE_STAGE_FRAGMENT_SHADER_BIT; |
| } else if (VK_IMAGE_LAYOUT_PREINITIALIZED == layout) { |
| return VK_PIPELINE_STAGE_HOST_BIT; |
| } else if (VK_IMAGE_LAYOUT_PRESENT_SRC_KHR == layout) { |
| return VK_PIPELINE_STAGE_COLOR_ATTACHMENT_OUTPUT_BIT; |
| } |
| |
| SkASSERT(VK_IMAGE_LAYOUT_UNDEFINED == layout); |
| return VK_PIPELINE_STAGE_TOP_OF_PIPE_BIT; |
| } |
| |
| VkAccessFlags VulkanTexture::LayoutToSrcAccessMask(const VkImageLayout layout) { |
| // Currently we assume we will never being doing any explict shader writes (this doesn't include |
| // color attachment or depth/stencil writes). So we will ignore the |
| // VK_MEMORY_OUTPUT_SHADER_WRITE_BIT. |
| |
| // We can only directly access the host memory if we are in preinitialized or general layout, |
| // and the image is linear. However, device access to images written by the host happens after |
| // vkQueueSubmit, which implicitly makes host writes _visible_ to the device, i.e. |
| // VK_ACCESS_HOST_WRITE_BIT is unnecessary. Host data is made _available_ to the device via |
| // vkFlushMappedMemoryRanges. |
| VkAccessFlags flags = 0; |
| if (VK_IMAGE_LAYOUT_GENERAL == layout) { |
| flags = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT | |
| VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT | |
| VK_ACCESS_TRANSFER_WRITE_BIT; |
| } else if (VK_IMAGE_LAYOUT_COLOR_ATTACHMENT_OPTIMAL == layout) { |
| flags = VK_ACCESS_COLOR_ATTACHMENT_WRITE_BIT; |
| } else if (VK_IMAGE_LAYOUT_DEPTH_STENCIL_ATTACHMENT_OPTIMAL == layout) { |
| flags = VK_ACCESS_DEPTH_STENCIL_ATTACHMENT_WRITE_BIT; |
| } else if (VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL == layout) { |
| flags = VK_ACCESS_TRANSFER_WRITE_BIT; |
| } else if (VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL == layout || |
| VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL == layout || |
| VK_IMAGE_LAYOUT_PRESENT_SRC_KHR == layout || |
| VK_IMAGE_LAYOUT_PREINITIALIZED == layout) { |
| // There are no writes that need to be made available |
| flags = 0; |
| } |
| return flags; |
| } |
| |
| const VulkanImageView* VulkanTexture::getImageView(VulkanImageView::Usage usage) const { |
| for (int i = 0; i < fImageViews.size(); ++i) { |
| if (fImageViews[i]->usage() == usage) { |
| return fImageViews[i].get(); |
| } |
| } |
| |
| auto sharedContext = static_cast<const VulkanSharedContext*>(this->sharedContext()); |
| const auto& vkTexInfo = this->vulkanTextureInfo(); |
| int miplevels = vkTexInfo.fMipmapped == Mipmapped::kYes |
| ? SkMipmap::ComputeLevelCount(this->dimensions()) + 1 |
| : 1; |
| auto imageView = VulkanImageView::Make(sharedContext, |
| fImage, |
| vkTexInfo.fFormat, |
| usage, |
| miplevels, |
| fYcbcrConversion); |
| return fImageViews.push_back(std::move(imageView)).get(); |
| } |
| |
| bool VulkanTexture::supportsInputAttachmentUsage() const { |
| return (this->vulkanTextureInfo().fImageUsageFlags & VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT); |
| } |
| |
| size_t VulkanTexture::onUpdateGpuMemorySize() { |
| if (!uses_lazy_memory(fMemoryAlloc)) { |
| // We don't expect non-transient textures to change their size over time. |
| return this->gpuMemorySize(); |
| } |
| |
| auto sharedContext = static_cast<const VulkanSharedContext*>(this->sharedContext()); |
| VkDeviceSize committedMemory; |
| VULKAN_CALL(sharedContext->interface(), |
| GetDeviceMemoryCommitment(sharedContext->device(), |
| fMemoryAlloc.fMemory, |
| &committedMemory)); |
| return committedMemory; |
| } |
| |
| sk_sp<VulkanDescriptorSet> VulkanTexture::getCachedSingleTextureDescriptorSet( |
| const Sampler* sampler) const { |
| SkASSERT(sampler); |
| for (auto& cachedSet : fCachedSingleTextureDescSets) { |
| if (cachedSet.first->uniqueID() == sampler->uniqueID()) { |
| return cachedSet.second; |
| } |
| } |
| return nullptr; |
| } |
| |
| void VulkanTexture::addCachedSingleTextureDescriptorSet(sk_sp<VulkanDescriptorSet> set, |
| sk_sp<const Sampler> sampler) const { |
| SkASSERT(set); |
| SkASSERT(sampler); |
| fCachedSingleTextureDescSets.push_back(std::make_pair(std::move(sampler), std::move(set))); |
| } |
| |
| sk_sp<VulkanFramebuffer> VulkanTexture::getCachedFramebuffer( |
| const RenderPassDesc& renderPassDesc, |
| const VulkanTexture* msaaTexture, |
| const VulkanTexture* depthStencilTexture) const { |
| for (auto& cachedFB : fCachedFramebuffers) { |
| if (cachedFB->compatible(renderPassDesc, msaaTexture, depthStencilTexture)) { |
| return cachedFB; |
| } |
| } |
| return nullptr; |
| } |
| |
| void VulkanTexture::addCachedFramebuffer(sk_sp<VulkanFramebuffer> fb) { |
| SkASSERT(fb); |
| fCachedFramebuffers.push_back(std::move(fb)); |
| } |
| |
| bool VulkanTexture::canUploadOnHost(const UploadSource& source) const { |
| // Can't use host-image-copy if the usage flag is not set. |
| if ((this->vulkanTextureInfo().fImageUsageFlags & VK_IMAGE_USAGE_HOST_TRANSFER_BIT) == 0) { |
| return false; |
| } |
| |
| // Can't use host-image-copy if the image is busy on the GPU. |
| if (this->isTextureBusyOnGPU()) { |
| return false; |
| } |
| |
| if (source.isRGB888Format()) { |
| // Need to transform RGBX8 to RGBA8 in a temp memory anyway, might as well use the buffer |
| // upload path for faster temp memory -> image copy by the GPU. |
| return false; |
| } |
| |
| // For now, only use host-image-copy if the image has never been used. If needed in the future, |
| // we could inspect the VkPhysicalDeviceHostImageCopyProperties::pCopySrcLayouts array to know |
| // which layouts the image can be to be used with HIC. However, a better solution could be to |
| // recreate the VkImage even if the existing one is busy on the GPU, since this function |
| // entirely overwrites the texture anyway. |
| if (this->currentLayout() != VK_IMAGE_LAYOUT_UNDEFINED) { |
| return false; |
| } |
| |
| return true; |
| } |
| |
| bool VulkanTexture::uploadDataOnHost(const UploadSource& source, const SkIRect& dstRect) { |
| auto sharedContext = static_cast<const VulkanSharedContext*>(this->sharedContext()); |
| SkSpan<const MipLevel> levels = source.levels(); |
| const unsigned int mipLevelCount = levels.size(); |
| |
| const TextureInfo& textureInfo = this->textureInfo(); |
| const TextureFormat format = TextureInfoPriv::ViewFormat(textureInfo); |
| const VkImageAspectFlags aspectFlags = GetVkImageAspectFlags(format); |
| |
| SkASSERT(this->currentLayout() == VK_IMAGE_LAYOUT_UNDEFINED); |
| |
| VkHostImageLayoutTransitionInfo transition = {}; |
| transition.sType = VK_STRUCTURE_TYPE_HOST_IMAGE_LAYOUT_TRANSITION_INFO; |
| transition.image = fImage; |
| transition.oldLayout = VK_IMAGE_LAYOUT_UNDEFINED; |
| transition.newLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL; |
| transition.subresourceRange.aspectMask = aspectFlags; |
| transition.subresourceRange.levelCount = mipLevelCount; |
| transition.subresourceRange.layerCount = 1; |
| |
| if (VULKAN_CALL(sharedContext->interface(), |
| TransitionImageLayout(sharedContext->device(), 1, &transition)) != VK_SUCCESS) { |
| return false; |
| } |
| this->updateImageLayout(VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL); |
| |
| TArray<VkMemoryToImageCopy> copyRegions(mipLevelCount); |
| |
| // The assumption is either that we have no mipmaps, or that our rect is the entire texture |
| SkASSERT(mipLevelCount == 1 || dstRect == SkIRect::MakeSize(this->dimensions())); |
| |
| // Copy data mip by mip. |
| const int32_t offsetX = dstRect.x(); |
| const int32_t offsetY = dstRect.y(); |
| int32_t currentWidth = dstRect.width(); |
| int32_t currentHeight = dstRect.height(); |
| |
| for (unsigned int currentMipLevel = 0; currentMipLevel < mipLevelCount; currentMipLevel++) { |
| // Upload data for compressed formats are fully packed. If this changes, the division by |
| // bytes-per-pixel should be adjusted for compressed formats. |
| SkASSERT(source.compression() == SkTextureCompressionType::kNone || |
| levels[currentMipLevel].fRowBytes == 0); |
| |
| VkMemoryToImageCopy copyRegion = {}; |
| copyRegion.sType = VK_STRUCTURE_TYPE_MEMORY_TO_IMAGE_COPY; |
| copyRegion.pHostPointer = levels[currentMipLevel].fPixels; |
| copyRegion.memoryRowLength = levels[currentMipLevel].fRowBytes / source.bytesPerPixel(); |
| copyRegion.memoryImageHeight = 0; // Tightly packed |
| copyRegion.imageSubresource.aspectMask = aspectFlags; |
| copyRegion.imageSubresource.mipLevel = currentMipLevel; |
| copyRegion.imageSubresource.layerCount = 1; |
| copyRegion.imageOffset.x = offsetX; |
| copyRegion.imageOffset.y = offsetY; |
| copyRegion.imageExtent.width = currentWidth; |
| copyRegion.imageExtent.height = currentHeight; |
| copyRegion.imageExtent.depth = 1; |
| |
| copyRegions.push_back(copyRegion); |
| |
| // Calculate the extent for the next mip. The offset does not need modification, since it's |
| // zero if mipLevelCount > 1, asserted before the loop. |
| currentWidth = std::max(1, currentWidth / 2); |
| currentHeight = std::max(1, currentHeight / 2); |
| } |
| |
| VkCopyMemoryToImageInfo copyInfo = {}; |
| copyInfo.sType = VK_STRUCTURE_TYPE_COPY_MEMORY_TO_IMAGE_INFO; |
| copyInfo.dstImage = fImage; |
| copyInfo.dstImageLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL; |
| copyInfo.regionCount = mipLevelCount; |
| copyInfo.pRegions = copyRegions.data(); |
| |
| const VkResult result = VULKAN_CALL(sharedContext->interface(), |
| CopyMemoryToImage(sharedContext->device(), ©Info)); |
| return result == VK_SUCCESS; |
| } |
| |
| } // namespace skgpu::graphite |