| /* |
| * Copyright 2017 Google Inc. |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| |
| #include <new> |
| |
| #include "include/core/SkPoint.h" |
| #include "include/core/SkPoint3.h" |
| #include "include/gpu/GrRecordingContext.h" |
| #include "include/private/base/SkFloatingPoint.h" |
| #include "include/private/base/SkTo.h" |
| #include "src/base/SkMathPriv.h" |
| #include "src/core/SkBlendModePriv.h" |
| #include "src/core/SkMatrixPriv.h" |
| #include "src/core/SkRectPriv.h" |
| #include "src/gpu/ganesh/GrAppliedClip.h" |
| #include "src/gpu/ganesh/GrCaps.h" |
| #include "src/gpu/ganesh/GrDrawOpTest.h" |
| #include "src/gpu/ganesh/GrGeometryProcessor.h" |
| #include "src/gpu/ganesh/GrGpu.h" |
| #include "src/gpu/ganesh/GrMemoryPool.h" |
| #include "src/gpu/ganesh/GrOpFlushState.h" |
| #include "src/gpu/ganesh/GrOpsTypes.h" |
| #include "src/gpu/ganesh/GrRecordingContextPriv.h" |
| #include "src/gpu/ganesh/GrResourceProvider.h" |
| #include "src/gpu/ganesh/GrResourceProviderPriv.h" |
| #include "src/gpu/ganesh/GrShaderCaps.h" |
| #include "src/gpu/ganesh/GrTexture.h" |
| #include "src/gpu/ganesh/GrTextureProxy.h" |
| #include "src/gpu/ganesh/SkGr.h" |
| #include "src/gpu/ganesh/SurfaceDrawContext.h" |
| #include "src/gpu/ganesh/effects/GrBlendFragmentProcessor.h" |
| #include "src/gpu/ganesh/effects/GrTextureEffect.h" |
| #include "src/gpu/ganesh/geometry/GrQuad.h" |
| #include "src/gpu/ganesh/geometry/GrQuadBuffer.h" |
| #include "src/gpu/ganesh/geometry/GrQuadUtils.h" |
| #include "src/gpu/ganesh/geometry/GrRect.h" |
| #include "src/gpu/ganesh/glsl/GrGLSLVarying.h" |
| #include "src/gpu/ganesh/ops/FillRectOp.h" |
| #include "src/gpu/ganesh/ops/GrMeshDrawOp.h" |
| #include "src/gpu/ganesh/ops/GrSimpleMeshDrawOpHelper.h" |
| #include "src/gpu/ganesh/ops/QuadPerEdgeAA.h" |
| #include "src/gpu/ganesh/ops/TextureOp.h" |
| |
| #if GR_TEST_UTILS |
| #include "src/gpu/ganesh/GrProxyProvider.h" |
| #endif |
| |
| using namespace skgpu::ganesh; |
| |
| namespace { |
| |
| using Subset = skgpu::v1::QuadPerEdgeAA::Subset; |
| using VertexSpec = skgpu::v1::QuadPerEdgeAA::VertexSpec; |
| using ColorType = skgpu::v1::QuadPerEdgeAA::ColorType; |
| |
| // Extracts lengths of vertical and horizontal edges of axis-aligned quad. "width" is the edge |
| // between v0 and v2 (or v1 and v3), "height" is the edge between v0 and v1 (or v2 and v3). |
| SkSize axis_aligned_quad_size(const GrQuad& quad) { |
| SkASSERT(quad.quadType() == GrQuad::Type::kAxisAligned); |
| // Simplification of regular edge length equation, since it's axis aligned and can avoid sqrt |
| float dw = sk_float_abs(quad.x(2) - quad.x(0)) + sk_float_abs(quad.y(2) - quad.y(0)); |
| float dh = sk_float_abs(quad.x(1) - quad.x(0)) + sk_float_abs(quad.y(1) - quad.y(0)); |
| return {dw, dh}; |
| } |
| |
| std::tuple<bool /* filter */, |
| bool /* mipmap */> |
| filter_and_mm_have_effect(const GrQuad& srcQuad, const GrQuad& dstQuad) { |
| // If not axis-aligned in src or dst, then always say it has an effect |
| if (srcQuad.quadType() != GrQuad::Type::kAxisAligned || |
| dstQuad.quadType() != GrQuad::Type::kAxisAligned) { |
| return {true, true}; |
| } |
| |
| SkRect srcRect; |
| SkRect dstRect; |
| if (srcQuad.asRect(&srcRect) && dstQuad.asRect(&dstRect)) { |
| // Disable filtering when there is no scaling (width and height are the same), and the |
| // top-left corners have the same fraction (so src and dst snap to the pixel grid |
| // identically). |
| SkASSERT(srcRect.isSorted()); |
| bool filter = srcRect.width() != dstRect.width() || srcRect.height() != dstRect.height() || |
| SkScalarFraction(srcRect.fLeft) != SkScalarFraction(dstRect.fLeft) || |
| SkScalarFraction(srcRect.fTop) != SkScalarFraction(dstRect.fTop); |
| bool mm = srcRect.width() > dstRect.width() || srcRect.height() > dstRect.height(); |
| return {filter, mm}; |
| } |
| // Extract edge lengths |
| SkSize srcSize = axis_aligned_quad_size(srcQuad); |
| SkSize dstSize = axis_aligned_quad_size(dstQuad); |
| // Although the quads are axis-aligned, the local coordinate system is transformed such |
| // that fractionally-aligned sample centers will not align with the device coordinate system |
| // So disable filtering when edges are the same length and both srcQuad and dstQuad |
| // 0th vertex is integer aligned. |
| bool filter = srcSize != dstSize || |
| !SkScalarIsInt(srcQuad.x(0)) || |
| !SkScalarIsInt(srcQuad.y(0)) || |
| !SkScalarIsInt(dstQuad.x(0)) || |
| !SkScalarIsInt(dstQuad.y(0)); |
| bool mm = srcSize.fWidth > dstSize.fWidth || srcSize.fHeight > dstSize.fHeight; |
| return {filter, mm}; |
| } |
| |
| // Describes function for normalizing src coords: [x * iw, y * ih + yOffset] can represent |
| // regular and rectangular textures, w/ or w/o origin correction. |
| struct NormalizationParams { |
| float fIW; // 1 / width of texture, or 1.0 for texture rectangles |
| float fInvH; // 1 / height of texture, or 1.0 for tex rects, X -1 if bottom-left origin |
| float fYOffset; // 0 for top-left origin, height of [normalized] tex if bottom-left |
| }; |
| NormalizationParams proxy_normalization_params(const GrSurfaceProxy* proxy, |
| GrSurfaceOrigin origin) { |
| // Whether or not the proxy is instantiated, this is the size its texture will be, so we can |
| // normalize the src coordinates up front. |
| SkISize dimensions = proxy->backingStoreDimensions(); |
| float iw, ih, h; |
| if (proxy->backendFormat().textureType() == GrTextureType::kRectangle) { |
| iw = ih = 1.f; |
| h = dimensions.height(); |
| } else { |
| iw = 1.f / dimensions.width(); |
| ih = 1.f / dimensions.height(); |
| h = 1.f; |
| } |
| |
| if (origin == kBottomLeft_GrSurfaceOrigin) { |
| return {iw, -ih, h}; |
| } else { |
| return {iw, ih, 0.0f}; |
| } |
| } |
| |
| // Normalize the subset. If 'subsetRect' is null, it is assumed no subset constraint is desired, |
| // so a sufficiently large rect is returned even if the quad ends up batched with an op that uses |
| // subsets overall. When there is a subset it will be inset based on the filter mode. Normalization |
| // and y-flipping are applied as indicated by NormalizationParams. |
| SkRect normalize_and_inset_subset(GrSamplerState::Filter filter, |
| const NormalizationParams& params, |
| const SkRect* subsetRect) { |
| static constexpr SkRect kLargeRect = {-100000, -100000, 1000000, 1000000}; |
| if (!subsetRect) { |
| // Either the quad has no subset constraint and is batched with a subset constrained op |
| // (in which case we want a subset that doesn't restrict normalized tex coords), or the |
| // entire op doesn't use the subset, in which case the returned value is ignored. |
| return kLargeRect; |
| } |
| |
| auto ltrb = skvx::Vec<4, float>::Load(subsetRect); |
| auto flipHi = skvx::Vec<4, float>({1.f, 1.f, -1.f, -1.f}); |
| if (filter == GrSamplerState::Filter::kNearest) { |
| // Make sure our insetting puts us at pixel centers. |
| ltrb = skvx::floor(ltrb*flipHi)*flipHi; |
| } |
| // Inset with pin to the rect center. |
| ltrb += skvx::Vec<4, float>({.5f, .5f, -.5f, -.5f}); |
| auto mid = (skvx::shuffle<2, 3, 0, 1>(ltrb) + ltrb)*0.5f; |
| ltrb = skvx::min(ltrb*flipHi, mid*flipHi)*flipHi; |
| |
| // Normalize and offset |
| ltrb = ltrb * skvx::Vec<4, float>{params.fIW, params.fInvH, params.fIW, params.fInvH} + |
| skvx::Vec<4, float>{0.f, params.fYOffset, 0.f, params.fYOffset}; |
| if (params.fInvH < 0.f) { |
| // Flip top and bottom to keep the rect sorted when loaded back to SkRect. |
| ltrb = skvx::shuffle<0, 3, 2, 1>(ltrb); |
| } |
| |
| SkRect out; |
| ltrb.store(&out); |
| return out; |
| } |
| |
| // Normalizes logical src coords and corrects for origin |
| void normalize_src_quad(const NormalizationParams& params, |
| GrQuad* srcQuad) { |
| // The src quad should not have any perspective |
| SkASSERT(!srcQuad->hasPerspective()); |
| skvx::Vec<4, float> xs = srcQuad->x4f() * params.fIW; |
| skvx::Vec<4, float> ys = srcQuad->y4f() * params.fInvH + params.fYOffset; |
| xs.store(srcQuad->xs()); |
| ys.store(srcQuad->ys()); |
| } |
| |
| // Count the number of proxy runs in the entry set. This usually is already computed by |
| // SkGpuDevice, but when the BatchLengthLimiter chops the set up it must determine a new proxy count |
| // for each split. |
| int proxy_run_count(const GrTextureSetEntry set[], int count) { |
| int actualProxyRunCount = 0; |
| const GrSurfaceProxy* lastProxy = nullptr; |
| for (int i = 0; i < count; ++i) { |
| if (set[i].fProxyView.proxy() != lastProxy) { |
| actualProxyRunCount++; |
| lastProxy = set[i].fProxyView.proxy(); |
| } |
| } |
| return actualProxyRunCount; |
| } |
| |
| bool safe_to_ignore_subset_rect(GrAAType aaType, GrSamplerState::Filter filter, |
| const DrawQuad& quad, const SkRect& subsetRect) { |
| // If both the device and local quad are both axis-aligned, and filtering is off, the local quad |
| // can push all the way up to the edges of the the subset rect and the sampler shouldn't |
| // overshoot. Unfortunately, antialiasing adds enough jitter that we can only rely on this in |
| // the non-antialiased case. |
| SkRect localBounds = quad.fLocal.bounds(); |
| if (aaType == GrAAType::kNone && |
| filter == GrSamplerState::Filter::kNearest && |
| quad.fDevice.quadType() == GrQuad::Type::kAxisAligned && |
| quad.fLocal.quadType() == GrQuad::Type::kAxisAligned && |
| subsetRect.contains(localBounds)) { |
| |
| return true; |
| } |
| |
| // If the local quad is inset by at least 0.5 pixels into the subset rect's bounds, the |
| // sampler shouldn't overshoot, even when antialiasing and filtering is taken into account. |
| if (subsetRect.makeInset(0.5f, 0.5f).contains(localBounds)) { |
| return true; |
| } |
| |
| // The subset rect cannot be ignored safely. |
| return false; |
| } |
| |
| /** |
| * Op that implements TextureOp::Make. It draws textured quads. Each quad can modulate against a |
| * the texture by color. The blend with the destination is always src-over. The edges are non-AA. |
| */ |
| class TextureOpImpl final : public GrMeshDrawOp { |
| public: |
| using Saturate = TextureOp::Saturate; |
| |
| static GrOp::Owner Make(GrRecordingContext* context, |
| GrSurfaceProxyView proxyView, |
| sk_sp<GrColorSpaceXform> textureXform, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| const SkPMColor4f& color, |
| Saturate saturate, |
| GrAAType aaType, |
| DrawQuad* quad, |
| const SkRect* subset) { |
| |
| return GrOp::Make<TextureOpImpl>(context, std::move(proxyView), std::move(textureXform), |
| filter, mm, color, saturate, aaType, quad, subset); |
| } |
| |
| static GrOp::Owner Make(GrRecordingContext* context, |
| GrTextureSetEntry set[], |
| int cnt, |
| int proxyRunCnt, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| Saturate saturate, |
| GrAAType aaType, |
| SkCanvas::SrcRectConstraint constraint, |
| const SkMatrix& viewMatrix, |
| sk_sp<GrColorSpaceXform> textureColorSpaceXform) { |
| // Allocate size based on proxyRunCnt, since that determines number of ViewCountPairs. |
| SkASSERT(proxyRunCnt <= cnt); |
| return GrOp::MakeWithExtraMemory<TextureOpImpl>( |
| context, sizeof(ViewCountPair) * (proxyRunCnt - 1), |
| set, cnt, proxyRunCnt, filter, mm, saturate, aaType, constraint, |
| viewMatrix, std::move(textureColorSpaceXform)); |
| } |
| |
| ~TextureOpImpl() override { |
| for (unsigned p = 1; p < fMetadata.fProxyCount; ++p) { |
| fViewCountPairs[p].~ViewCountPair(); |
| } |
| } |
| |
| const char* name() const override { return "TextureOp"; } |
| |
| void visitProxies(const GrVisitProxyFunc& func) const override { |
| bool mipped = (fMetadata.mipmapMode() != GrSamplerState::MipmapMode::kNone); |
| for (unsigned p = 0; p < fMetadata.fProxyCount; ++p) { |
| func(fViewCountPairs[p].fProxy.get(), GrMipmapped(mipped)); |
| } |
| if (fDesc && fDesc->fProgramInfo) { |
| fDesc->fProgramInfo->visitFPProxies(func); |
| } |
| } |
| |
| #ifdef SK_DEBUG |
| static void ValidateResourceLimits() { |
| // The op implementation has an upper bound on the number of quads that it can represent. |
| // However, the resource manager imposes its own limit on the number of quads, which should |
| // always be lower than the numerical limit this op can hold. |
| using CountStorage = decltype(Metadata::fTotalQuadCount); |
| CountStorage maxQuadCount = std::numeric_limits<CountStorage>::max(); |
| // GrResourceProvider::Max...() is typed as int, so don't compare across signed/unsigned. |
| int resourceLimit = SkTo<int>(maxQuadCount); |
| SkASSERT(GrResourceProvider::MaxNumAAQuads() <= resourceLimit && |
| GrResourceProvider::MaxNumNonAAQuads() <= resourceLimit); |
| } |
| #endif |
| |
| GrProcessorSet::Analysis finalize(const GrCaps& caps, const GrAppliedClip*, |
| GrClampType clampType) override { |
| SkASSERT(fMetadata.colorType() == ColorType::kNone); |
| auto iter = fQuads.metadata(); |
| while(iter.next()) { |
| auto colorType = skgpu::v1::QuadPerEdgeAA::MinColorType(iter->fColor); |
| colorType = std::max(static_cast<ColorType>(fMetadata.fColorType), |
| colorType); |
| if (caps.reducedShaderMode()) { |
| colorType = std::max(colorType, ColorType::kByte); |
| } |
| fMetadata.fColorType = static_cast<uint16_t>(colorType); |
| } |
| return GrProcessorSet::EmptySetAnalysis(); |
| } |
| |
| FixedFunctionFlags fixedFunctionFlags() const override { |
| return fMetadata.aaType() == GrAAType::kMSAA ? FixedFunctionFlags::kUsesHWAA |
| : FixedFunctionFlags::kNone; |
| } |
| |
| DEFINE_OP_CLASS_ID |
| |
| private: |
| friend class ::GrOp; |
| |
| struct ColorSubsetAndAA { |
| ColorSubsetAndAA(const SkPMColor4f& color, const SkRect& subsetRect, GrQuadAAFlags aaFlags) |
| : fColor(color) |
| , fSubsetRect(subsetRect) |
| , fAAFlags(static_cast<uint16_t>(aaFlags)) { |
| SkASSERT(fAAFlags == static_cast<uint16_t>(aaFlags)); |
| } |
| |
| SkPMColor4f fColor; |
| // If the op doesn't use subsets, this is ignored. If the op uses subsets and the specific |
| // entry does not, this rect will equal kLargeRect, so it automatically has no effect. |
| SkRect fSubsetRect; |
| unsigned fAAFlags : 4; |
| |
| GrQuadAAFlags aaFlags() const { return static_cast<GrQuadAAFlags>(fAAFlags); } |
| }; |
| |
| struct ViewCountPair { |
| // Normally this would be a GrSurfaceProxyView, but TextureOp applies the GrOrigin right |
| // away so it doesn't need to be stored, and all ViewCountPairs in an op have the same |
| // swizzle so that is stored in the op metadata. |
| sk_sp<GrSurfaceProxy> fProxy; |
| int fQuadCnt; |
| }; |
| |
| // TextureOp and ViewCountPair are 8 byte aligned. This is packed into 8 bytes to minimally |
| // increase the size of the op; increasing the op size can have a surprising impact on |
| // performance (since texture ops are one of the most commonly used in an app). |
| struct Metadata { |
| // AAType must be filled after initialization; ColorType is determined in finalize() |
| Metadata(const skgpu::Swizzle& swizzle, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| Subset subset, |
| Saturate saturate) |
| : fSwizzle(swizzle) |
| , fProxyCount(1) |
| , fTotalQuadCount(1) |
| , fFilter(static_cast<uint16_t>(filter)) |
| , fMipmapMode(static_cast<uint16_t>(mm)) |
| , fAAType(static_cast<uint16_t>(GrAAType::kNone)) |
| , fColorType(static_cast<uint16_t>(ColorType::kNone)) |
| , fSubset(static_cast<uint16_t>(subset)) |
| , fSaturate(static_cast<uint16_t>(saturate)) {} |
| |
| skgpu::Swizzle fSwizzle; // sizeof(skgpu::Swizzle) == uint16_t |
| uint16_t fProxyCount; |
| // This will be >= fProxyCount, since a proxy may be drawn multiple times |
| uint16_t fTotalQuadCount; |
| |
| // These must be based on uint16_t to help MSVC's pack bitfields optimally |
| uint16_t fFilter : 2; // GrSamplerState::Filter |
| uint16_t fMipmapMode : 2; // GrSamplerState::MipmapMode |
| uint16_t fAAType : 2; // GrAAType |
| uint16_t fColorType : 2; // GrQuadPerEdgeAA::ColorType |
| uint16_t fSubset : 1; // bool |
| uint16_t fSaturate : 1; // bool |
| uint16_t fUnused : 6; // # of bits left before Metadata exceeds 8 bytes |
| |
| GrSamplerState::Filter filter() const { |
| return static_cast<GrSamplerState::Filter>(fFilter); |
| } |
| GrSamplerState::MipmapMode mipmapMode() const { |
| return static_cast<GrSamplerState::MipmapMode>(fMipmapMode); |
| } |
| GrAAType aaType() const { return static_cast<GrAAType>(fAAType); } |
| ColorType colorType() const { return static_cast<ColorType>(fColorType); } |
| Subset subset() const { return static_cast<Subset>(fSubset); } |
| Saturate saturate() const { return static_cast<Saturate>(fSaturate); } |
| |
| static_assert(GrSamplerState::kFilterCount <= 4); |
| static_assert(kGrAATypeCount <= 4); |
| static_assert(skgpu::v1::QuadPerEdgeAA::kColorTypeCount <= 4); |
| }; |
| static_assert(sizeof(Metadata) == 8); |
| |
| // This descriptor is used to store the draw info we decide on during on(Pre)PrepareDraws. We |
| // store the data in a separate struct in order to minimize the size of the TextureOp. |
| // Historically, increasing the TextureOp's size has caused surprising perf regressions, but we |
| // may want to re-evaluate whether this is still necessary. |
| // |
| // In the onPrePrepareDraws case it is allocated in the creation-time opData arena, and |
| // allocatePrePreparedVertices is also called. |
| // |
| // In the onPrepareDraws case this descriptor is allocated in the flush-time arena (i.e., as |
| // part of the flushState). |
| struct Desc { |
| VertexSpec fVertexSpec; |
| int fNumProxies = 0; |
| int fNumTotalQuads = 0; |
| |
| // This member variable is only used by 'onPrePrepareDraws'. |
| char* fPrePreparedVertices = nullptr; |
| |
| GrProgramInfo* fProgramInfo = nullptr; |
| |
| sk_sp<const GrBuffer> fIndexBuffer; |
| sk_sp<const GrBuffer> fVertexBuffer; |
| int fBaseVertex; |
| |
| // How big should 'fVertices' be to hold all the vertex data? |
| size_t totalSizeInBytes() const { |
| return this->totalNumVertices() * fVertexSpec.vertexSize(); |
| } |
| |
| int totalNumVertices() const { |
| return fNumTotalQuads * fVertexSpec.verticesPerQuad(); |
| } |
| |
| void allocatePrePreparedVertices(SkArenaAlloc* arena) { |
| fPrePreparedVertices = arena->makeArrayDefault<char>(this->totalSizeInBytes()); |
| } |
| }; |
| // If subsetRect is not null it will be used to apply a strict src rect-style constraint. |
| TextureOpImpl(GrSurfaceProxyView proxyView, |
| sk_sp<GrColorSpaceXform> textureColorSpaceXform, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| const SkPMColor4f& color, |
| Saturate saturate, |
| GrAAType aaType, |
| DrawQuad* quad, |
| const SkRect* subsetRect) |
| : INHERITED(ClassID()) |
| , fQuads(1, true /* includes locals */) |
| , fTextureColorSpaceXform(std::move(textureColorSpaceXform)) |
| , fDesc(nullptr) |
| , fMetadata(proxyView.swizzle(), filter, mm, Subset(!!subsetRect), saturate) { |
| // Clean up disparities between the overall aa type and edge configuration and apply |
| // optimizations based on the rect and matrix when appropriate |
| GrQuadUtils::ResolveAAType(aaType, quad->fEdgeFlags, quad->fDevice, |
| &aaType, &quad->fEdgeFlags); |
| fMetadata.fAAType = static_cast<uint16_t>(aaType); |
| |
| // We expect our caller to have already caught this optimization. |
| SkASSERT(!subsetRect || |
| !subsetRect->contains(proxyView.proxy()->backingStoreBoundsRect())); |
| |
| // We may have had a strict constraint with nearest filter solely due to possible AA bloat. |
| // Try to identify cases where the subsetting isn't actually necessary, and skip it. |
| if (subsetRect) { |
| if (safe_to_ignore_subset_rect(aaType, filter, *quad, *subsetRect)) { |
| subsetRect = nullptr; |
| fMetadata.fSubset = static_cast<uint16_t>(Subset::kNo); |
| } |
| } |
| |
| // Normalize src coordinates and the subset (if set) |
| NormalizationParams params = proxy_normalization_params(proxyView.proxy(), |
| proxyView.origin()); |
| normalize_src_quad(params, &quad->fLocal); |
| SkRect subset = normalize_and_inset_subset(filter, params, subsetRect); |
| |
| // Set bounds before clipping so we don't have to worry about unioning the bounds of |
| // the two potential quads (GrQuad::bounds() is perspective-safe). |
| bool hairline = GrQuadUtils::WillUseHairline(quad->fDevice, aaType, quad->fEdgeFlags); |
| this->setBounds(quad->fDevice.bounds(), HasAABloat(aaType == GrAAType::kCoverage), |
| hairline ? IsHairline::kYes : IsHairline::kNo); |
| int quadCount = this->appendQuad(quad, color, subset); |
| fViewCountPairs[0] = {proxyView.detachProxy(), quadCount}; |
| } |
| |
| TextureOpImpl(GrTextureSetEntry set[], |
| int cnt, |
| int proxyRunCnt, |
| const GrSamplerState::Filter filter, |
| const GrSamplerState::MipmapMode mm, |
| const Saturate saturate, |
| const GrAAType aaType, |
| const SkCanvas::SrcRectConstraint constraint, |
| const SkMatrix& viewMatrix, |
| sk_sp<GrColorSpaceXform> textureColorSpaceXform) |
| : INHERITED(ClassID()) |
| , fQuads(cnt, true /* includes locals */) |
| , fTextureColorSpaceXform(std::move(textureColorSpaceXform)) |
| , fDesc(nullptr) |
| , fMetadata(set[0].fProxyView.swizzle(), |
| GrSamplerState::Filter::kNearest, |
| GrSamplerState::MipmapMode::kNone, |
| Subset::kNo, |
| saturate) { |
| // Update counts to reflect the batch op |
| fMetadata.fProxyCount = SkToUInt(proxyRunCnt); |
| fMetadata.fTotalQuadCount = SkToUInt(cnt); |
| |
| SkRect bounds = SkRectPriv::MakeLargestInverted(); |
| |
| GrAAType netAAType = GrAAType::kNone; // aa type maximally compatible with all dst rects |
| Subset netSubset = Subset::kNo; |
| GrSamplerState::Filter netFilter = GrSamplerState::Filter::kNearest; |
| GrSamplerState::MipmapMode netMM = GrSamplerState::MipmapMode::kNone; |
| bool hasSubpixel = false; |
| |
| const GrSurfaceProxy* curProxy = nullptr; |
| |
| // 'q' is the index in 'set' and fQuadBuffer; 'p' is the index in fViewCountPairs and only |
| // increases when set[q]'s proxy changes. |
| int p = 0; |
| for (int q = 0; q < cnt; ++q) { |
| SkASSERT(mm == GrSamplerState::MipmapMode::kNone || |
| (set[0].fProxyView.proxy()->asTextureProxy()->mipmapped() == |
| GrMipmapped::kYes)); |
| if (q == 0) { |
| // We do not placement new the first ViewCountPair since that one is allocated and |
| // initialized as part of the TextureOp creation. |
| fViewCountPairs[0].fProxy = set[0].fProxyView.detachProxy(); |
| fViewCountPairs[0].fQuadCnt = 0; |
| curProxy = fViewCountPairs[0].fProxy.get(); |
| } else if (set[q].fProxyView.proxy() != curProxy) { |
| // We must placement new the ViewCountPairs here so that the sk_sps in the |
| // GrSurfaceProxyView get initialized properly. |
| new(&fViewCountPairs[++p])ViewCountPair({set[q].fProxyView.detachProxy(), 0}); |
| |
| curProxy = fViewCountPairs[p].fProxy.get(); |
| SkASSERT(GrTextureProxy::ProxiesAreCompatibleAsDynamicState( |
| curProxy, fViewCountPairs[0].fProxy.get())); |
| SkASSERT(fMetadata.fSwizzle == set[q].fProxyView.swizzle()); |
| } // else another quad referencing the same proxy |
| |
| SkMatrix ctm = viewMatrix; |
| if (set[q].fPreViewMatrix) { |
| ctm.preConcat(*set[q].fPreViewMatrix); |
| } |
| |
| // Use dstRect/srcRect unless dstClip is provided, in which case derive new source |
| // coordinates by mapping dstClipQuad by the dstRect to srcRect transform. |
| DrawQuad quad; |
| if (set[q].fDstClipQuad) { |
| quad.fDevice = GrQuad::MakeFromSkQuad(set[q].fDstClipQuad, ctm); |
| |
| SkPoint srcPts[4]; |
| GrMapRectPoints(set[q].fDstRect, set[q].fSrcRect, set[q].fDstClipQuad, srcPts, 4); |
| quad.fLocal = GrQuad::MakeFromSkQuad(srcPts, SkMatrix::I()); |
| } else { |
| quad.fDevice = GrQuad::MakeFromRect(set[q].fDstRect, ctm); |
| quad.fLocal = GrQuad(set[q].fSrcRect); |
| } |
| |
| // This may be reduced per-quad from the requested aggregate filtering level, and used |
| // to determine if the subset is needed for the entry as well. |
| GrSamplerState::Filter filterForQuad = filter; |
| if (netFilter != filter || netMM != mm) { |
| // The only way netFilter != filter is if linear is requested and we haven't yet |
| // found a quad that requires linear (so net is still nearest). Similar for mip |
| // mapping. |
| SkASSERT(filter == netFilter || |
| (netFilter == GrSamplerState::Filter::kNearest && filter > netFilter)); |
| SkASSERT(mm == netMM || |
| (netMM == GrSamplerState::MipmapMode::kNone && mm > netMM)); |
| auto [mustFilter, mustMM] = filter_and_mm_have_effect(quad.fLocal, quad.fDevice); |
| if (filter != GrSamplerState::Filter::kNearest) { |
| if (mustFilter) { |
| netFilter = filter; // upgrade batch to higher filter level |
| } else { |
| filterForQuad = GrSamplerState::Filter::kNearest; // downgrade entry |
| } |
| } |
| if (mustMM && mm != GrSamplerState::MipmapMode::kNone) { |
| netMM = mm; |
| } |
| } |
| |
| // Determine the AA type for the quad, then merge with net AA type |
| GrAAType aaForQuad; |
| GrQuadUtils::ResolveAAType(aaType, set[q].fAAFlags, quad.fDevice, |
| &aaForQuad, &quad.fEdgeFlags); |
| // Update overall bounds of the op as the union of all quads |
| bounds.joinPossiblyEmptyRect(quad.fDevice.bounds()); |
| hasSubpixel |= GrQuadUtils::WillUseHairline(quad.fDevice, aaForQuad, quad.fEdgeFlags); |
| |
| // Resolve sets aaForQuad to aaType or None, there is never a change between aa methods |
| SkASSERT(aaForQuad == GrAAType::kNone || aaForQuad == aaType); |
| if (netAAType == GrAAType::kNone && aaForQuad != GrAAType::kNone) { |
| netAAType = aaType; |
| } |
| |
| // Calculate metadata for the entry |
| const SkRect* subsetForQuad = nullptr; |
| if (constraint == SkCanvas::kStrict_SrcRectConstraint) { |
| // Check (briefly) if the subset rect is actually needed for this set entry. |
| SkRect* subsetRect = &set[q].fSrcRect; |
| if (!subsetRect->contains(curProxy->backingStoreBoundsRect())) { |
| if (!safe_to_ignore_subset_rect(aaForQuad, filterForQuad, quad, *subsetRect)) { |
| netSubset = Subset::kYes; |
| subsetForQuad = subsetRect; |
| } |
| } |
| } |
| |
| // Normalize the src quads and apply origin |
| NormalizationParams proxyParams = proxy_normalization_params( |
| curProxy, set[q].fProxyView.origin()); |
| normalize_src_quad(proxyParams, &quad.fLocal); |
| |
| // This subset may represent a no-op, otherwise it will have the origin and dimensions |
| // of the texture applied to it. |
| SkRect subset = normalize_and_inset_subset(filter, proxyParams, subsetForQuad); |
| |
| // Always append a quad (or 2 if perspective clipped), it just may refer back to a prior |
| // ViewCountPair (this frequently happens when Chrome draws 9-patches). |
| fViewCountPairs[p].fQuadCnt += this->appendQuad(&quad, set[q].fColor, subset); |
| } |
| // The # of proxy switches should match what was provided (+1 because we incremented p |
| // when a new proxy was encountered). |
| SkASSERT((p + 1) == fMetadata.fProxyCount); |
| SkASSERT(fQuads.count() == fMetadata.fTotalQuadCount); |
| |
| fMetadata.fAAType = static_cast<uint16_t>(netAAType); |
| fMetadata.fFilter = static_cast<uint16_t>(netFilter); |
| fMetadata.fSubset = static_cast<uint16_t>(netSubset); |
| |
| this->setBounds(bounds, HasAABloat(netAAType == GrAAType::kCoverage), |
| hasSubpixel ? IsHairline::kYes : IsHairline::kNo); |
| } |
| |
| int appendQuad(DrawQuad* quad, const SkPMColor4f& color, const SkRect& subset) { |
| DrawQuad extra; |
| // Always clip to W0 to stay consistent with GrQuad::bounds |
| int quadCount = GrQuadUtils::ClipToW0(quad, &extra); |
| if (quadCount == 0) { |
| // We can't discard the op at this point, but disable AA flags so it won't go through |
| // inset/outset processing |
| quad->fEdgeFlags = GrQuadAAFlags::kNone; |
| quadCount = 1; |
| } |
| fQuads.append(quad->fDevice, {color, subset, quad->fEdgeFlags}, &quad->fLocal); |
| if (quadCount > 1) { |
| fQuads.append(extra.fDevice, {color, subset, extra.fEdgeFlags}, &extra.fLocal); |
| fMetadata.fTotalQuadCount++; |
| } |
| return quadCount; |
| } |
| |
| GrProgramInfo* programInfo() override { |
| // Although this Op implements its own onPrePrepareDraws it calls GrMeshDrawOps' version so |
| // this entry point will be called. |
| return (fDesc) ? fDesc->fProgramInfo : nullptr; |
| } |
| |
| void onCreateProgramInfo(const GrCaps* caps, |
| SkArenaAlloc* arena, |
| const GrSurfaceProxyView& writeView, |
| bool usesMSAASurface, |
| GrAppliedClip&& appliedClip, |
| const GrDstProxyView& dstProxyView, |
| GrXferBarrierFlags renderPassXferBarriers, |
| GrLoadOp colorLoadOp) override { |
| SkASSERT(fDesc); |
| |
| GrGeometryProcessor* gp; |
| |
| { |
| const GrBackendFormat& backendFormat = |
| fViewCountPairs[0].fProxy->backendFormat(); |
| |
| GrSamplerState samplerState = GrSamplerState(GrSamplerState::WrapMode::kClamp, |
| fMetadata.filter()); |
| |
| gp = skgpu::v1::QuadPerEdgeAA::MakeTexturedProcessor( |
| arena, fDesc->fVertexSpec, *caps->shaderCaps(), backendFormat, samplerState, |
| fMetadata.fSwizzle, std::move(fTextureColorSpaceXform), fMetadata.saturate()); |
| |
| SkASSERT(fDesc->fVertexSpec.vertexSize() == gp->vertexStride()); |
| } |
| |
| fDesc->fProgramInfo = GrSimpleMeshDrawOpHelper::CreateProgramInfo( |
| caps, arena, writeView, usesMSAASurface, std::move(appliedClip), dstProxyView, gp, |
| GrProcessorSet::MakeEmptySet(), fDesc->fVertexSpec.primitiveType(), |
| renderPassXferBarriers, colorLoadOp, GrPipeline::InputFlags::kNone); |
| } |
| |
| void onPrePrepareDraws(GrRecordingContext* context, |
| const GrSurfaceProxyView& writeView, |
| GrAppliedClip* clip, |
| const GrDstProxyView& dstProxyView, |
| GrXferBarrierFlags renderPassXferBarriers, |
| GrLoadOp colorLoadOp) override { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| |
| SkDEBUGCODE(this->validate();) |
| SkASSERT(!fDesc); |
| |
| SkArenaAlloc* arena = context->priv().recordTimeAllocator(); |
| |
| fDesc = arena->make<Desc>(); |
| this->characterize(fDesc); |
| fDesc->allocatePrePreparedVertices(arena); |
| FillInVertices(*context->priv().caps(), this, fDesc, fDesc->fPrePreparedVertices); |
| |
| // This will call onCreateProgramInfo and register the created program with the DDL. |
| this->INHERITED::onPrePrepareDraws(context, writeView, clip, dstProxyView, |
| renderPassXferBarriers, colorLoadOp); |
| } |
| |
| static void FillInVertices(const GrCaps& caps, |
| TextureOpImpl* texOp, |
| Desc* desc, |
| char* vertexData) { |
| SkASSERT(vertexData); |
| |
| SkDEBUGCODE(int totQuadsSeen = 0;) |
| SkDEBUGCODE(int totVerticesSeen = 0;) |
| SkDEBUGCODE(const size_t vertexSize = desc->fVertexSpec.vertexSize();) |
| SkDEBUGCODE(auto startMark{vertexData};) |
| |
| skgpu::v1::QuadPerEdgeAA::Tessellator tessellator(desc->fVertexSpec, vertexData); |
| for (const auto& op : ChainRange<TextureOpImpl>(texOp)) { |
| auto iter = op.fQuads.iterator(); |
| for (unsigned p = 0; p < op.fMetadata.fProxyCount; ++p) { |
| const int quadCnt = op.fViewCountPairs[p].fQuadCnt; |
| SkDEBUGCODE(int meshVertexCnt = quadCnt * desc->fVertexSpec.verticesPerQuad()); |
| |
| for (int i = 0; i < quadCnt && iter.next(); ++i) { |
| SkASSERT(iter.isLocalValid()); |
| const ColorSubsetAndAA& info = iter.metadata(); |
| |
| tessellator.append(iter.deviceQuad(), iter.localQuad(), info.fColor, |
| info.fSubsetRect, info.aaFlags()); |
| } |
| |
| SkASSERT((totVerticesSeen + meshVertexCnt) * vertexSize |
| == (size_t)(tessellator.vertexMark() - startMark)); |
| |
| SkDEBUGCODE(totQuadsSeen += quadCnt;) |
| SkDEBUGCODE(totVerticesSeen += meshVertexCnt); |
| SkASSERT(totQuadsSeen * desc->fVertexSpec.verticesPerQuad() == totVerticesSeen); |
| } |
| |
| // If quad counts per proxy were calculated correctly, the entire iterator |
| // should have been consumed. |
| SkASSERT(!iter.next()); |
| } |
| |
| SkASSERT(desc->totalSizeInBytes() == (size_t)(tessellator.vertexMark() - startMark)); |
| SkASSERT(totQuadsSeen == desc->fNumTotalQuads); |
| SkASSERT(totVerticesSeen == desc->totalNumVertices()); |
| } |
| |
| #ifdef SK_DEBUG |
| static int validate_op(GrTextureType textureType, |
| GrAAType aaType, |
| skgpu::Swizzle swizzle, |
| const TextureOpImpl* op) { |
| SkASSERT(op->fMetadata.fSwizzle == swizzle); |
| |
| int quadCount = 0; |
| for (unsigned p = 0; p < op->fMetadata.fProxyCount; ++p) { |
| auto* proxy = op->fViewCountPairs[p].fProxy->asTextureProxy(); |
| quadCount += op->fViewCountPairs[p].fQuadCnt; |
| SkASSERT(proxy); |
| SkASSERT(proxy->textureType() == textureType); |
| } |
| |
| SkASSERT(aaType == op->fMetadata.aaType()); |
| return quadCount; |
| } |
| |
| void validate() const override { |
| // NOTE: Since this is debug-only code, we use the virtual asTextureProxy() |
| auto textureType = fViewCountPairs[0].fProxy->asTextureProxy()->textureType(); |
| GrAAType aaType = fMetadata.aaType(); |
| skgpu::Swizzle swizzle = fMetadata.fSwizzle; |
| |
| int quadCount = validate_op(textureType, aaType, swizzle, this); |
| |
| for (const GrOp* tmp = this->prevInChain(); tmp; tmp = tmp->prevInChain()) { |
| quadCount += validate_op(textureType, aaType, swizzle, |
| static_cast<const TextureOpImpl*>(tmp)); |
| } |
| |
| for (const GrOp* tmp = this->nextInChain(); tmp; tmp = tmp->nextInChain()) { |
| quadCount += validate_op(textureType, aaType, swizzle, |
| static_cast<const TextureOpImpl*>(tmp)); |
| } |
| |
| SkASSERT(quadCount == this->numChainedQuads()); |
| } |
| |
| #endif |
| |
| #if GR_TEST_UTILS |
| int numQuads() const final { return this->totNumQuads(); } |
| #endif |
| |
| void characterize(Desc* desc) const { |
| SkDEBUGCODE(this->validate();) |
| |
| GrQuad::Type quadType = GrQuad::Type::kAxisAligned; |
| ColorType colorType = ColorType::kNone; |
| GrQuad::Type srcQuadType = GrQuad::Type::kAxisAligned; |
| Subset subset = Subset::kNo; |
| GrAAType overallAAType = fMetadata.aaType(); |
| |
| desc->fNumProxies = 0; |
| desc->fNumTotalQuads = 0; |
| int maxQuadsPerMesh = 0; |
| |
| for (const auto& op : ChainRange<TextureOpImpl>(this)) { |
| if (op.fQuads.deviceQuadType() > quadType) { |
| quadType = op.fQuads.deviceQuadType(); |
| } |
| if (op.fQuads.localQuadType() > srcQuadType) { |
| srcQuadType = op.fQuads.localQuadType(); |
| } |
| if (op.fMetadata.subset() == Subset::kYes) { |
| subset = Subset::kYes; |
| } |
| colorType = std::max(colorType, op.fMetadata.colorType()); |
| desc->fNumProxies += op.fMetadata.fProxyCount; |
| |
| for (unsigned p = 0; p < op.fMetadata.fProxyCount; ++p) { |
| maxQuadsPerMesh = std::max(op.fViewCountPairs[p].fQuadCnt, maxQuadsPerMesh); |
| } |
| desc->fNumTotalQuads += op.totNumQuads(); |
| |
| if (op.fMetadata.aaType() == GrAAType::kCoverage) { |
| overallAAType = GrAAType::kCoverage; |
| } |
| } |
| |
| SkASSERT(desc->fNumTotalQuads == this->numChainedQuads()); |
| |
| SkASSERT(!CombinedQuadCountWillOverflow(overallAAType, false, desc->fNumTotalQuads)); |
| |
| auto indexBufferOption = skgpu::v1::QuadPerEdgeAA::CalcIndexBufferOption(overallAAType, |
| maxQuadsPerMesh); |
| |
| desc->fVertexSpec = VertexSpec(quadType, colorType, srcQuadType, /* hasLocal */ true, |
| subset, overallAAType, /* alpha as coverage */ true, |
| indexBufferOption); |
| |
| SkASSERT(desc->fNumTotalQuads <= skgpu::v1::QuadPerEdgeAA::QuadLimit(indexBufferOption)); |
| } |
| |
| int totNumQuads() const { |
| #ifdef SK_DEBUG |
| int tmp = 0; |
| for (unsigned p = 0; p < fMetadata.fProxyCount; ++p) { |
| tmp += fViewCountPairs[p].fQuadCnt; |
| } |
| SkASSERT(tmp == fMetadata.fTotalQuadCount); |
| #endif |
| |
| return fMetadata.fTotalQuadCount; |
| } |
| |
| int numChainedQuads() const { |
| int numChainedQuads = this->totNumQuads(); |
| |
| for (const GrOp* tmp = this->prevInChain(); tmp; tmp = tmp->prevInChain()) { |
| numChainedQuads += ((const TextureOpImpl*)tmp)->totNumQuads(); |
| } |
| |
| for (const GrOp* tmp = this->nextInChain(); tmp; tmp = tmp->nextInChain()) { |
| numChainedQuads += ((const TextureOpImpl*)tmp)->totNumQuads(); |
| } |
| |
| return numChainedQuads; |
| } |
| |
| // onPrePrepareDraws may or may not have been called at this point |
| void onPrepareDraws(GrMeshDrawTarget* target) override { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| |
| SkDEBUGCODE(this->validate();) |
| |
| SkASSERT(!fDesc || fDesc->fPrePreparedVertices); |
| |
| if (!fDesc) { |
| SkArenaAlloc* arena = target->allocator(); |
| fDesc = arena->make<Desc>(); |
| this->characterize(fDesc); |
| SkASSERT(!fDesc->fPrePreparedVertices); |
| } |
| |
| size_t vertexSize = fDesc->fVertexSpec.vertexSize(); |
| |
| void* vdata = target->makeVertexSpace(vertexSize, fDesc->totalNumVertices(), |
| &fDesc->fVertexBuffer, &fDesc->fBaseVertex); |
| if (!vdata) { |
| SkDebugf("Could not allocate vertices\n"); |
| return; |
| } |
| |
| if (fDesc->fVertexSpec.needsIndexBuffer()) { |
| fDesc->fIndexBuffer = skgpu::v1::QuadPerEdgeAA::GetIndexBuffer( |
| target, fDesc->fVertexSpec.indexBufferOption()); |
| if (!fDesc->fIndexBuffer) { |
| SkDebugf("Could not allocate indices\n"); |
| return; |
| } |
| } |
| |
| if (fDesc->fPrePreparedVertices) { |
| memcpy(vdata, fDesc->fPrePreparedVertices, fDesc->totalSizeInBytes()); |
| } else { |
| FillInVertices(target->caps(), this, fDesc, (char*) vdata); |
| } |
| } |
| |
| void onExecute(GrOpFlushState* flushState, const SkRect& chainBounds) override { |
| if (!fDesc->fVertexBuffer) { |
| return; |
| } |
| |
| if (fDesc->fVertexSpec.needsIndexBuffer() && !fDesc->fIndexBuffer) { |
| return; |
| } |
| |
| if (!fDesc->fProgramInfo) { |
| this->createProgramInfo(flushState); |
| SkASSERT(fDesc->fProgramInfo); |
| } |
| |
| flushState->bindPipelineAndScissorClip(*fDesc->fProgramInfo, chainBounds); |
| flushState->bindBuffers(std::move(fDesc->fIndexBuffer), nullptr, |
| std::move(fDesc->fVertexBuffer)); |
| |
| int totQuadsSeen = 0; |
| SkDEBUGCODE(int numDraws = 0;) |
| for (const auto& op : ChainRange<TextureOpImpl>(this)) { |
| for (unsigned p = 0; p < op.fMetadata.fProxyCount; ++p) { |
| const int quadCnt = op.fViewCountPairs[p].fQuadCnt; |
| SkASSERT(numDraws < fDesc->fNumProxies); |
| flushState->bindTextures(fDesc->fProgramInfo->geomProc(), |
| *op.fViewCountPairs[p].fProxy, |
| fDesc->fProgramInfo->pipeline()); |
| skgpu::v1::QuadPerEdgeAA::IssueDraw(flushState->caps(), flushState->opsRenderPass(), |
| fDesc->fVertexSpec, totQuadsSeen, quadCnt, |
| fDesc->totalNumVertices(), fDesc->fBaseVertex); |
| totQuadsSeen += quadCnt; |
| SkDEBUGCODE(++numDraws;) |
| } |
| } |
| |
| SkASSERT(totQuadsSeen == fDesc->fNumTotalQuads); |
| SkASSERT(numDraws == fDesc->fNumProxies); |
| } |
| |
| void propagateCoverageAAThroughoutChain() { |
| fMetadata.fAAType = static_cast<uint16_t>(GrAAType::kCoverage); |
| |
| for (GrOp* tmp = this->prevInChain(); tmp; tmp = tmp->prevInChain()) { |
| auto tex = static_cast<TextureOpImpl*>(tmp); |
| SkASSERT(tex->fMetadata.aaType() == GrAAType::kCoverage || |
| tex->fMetadata.aaType() == GrAAType::kNone); |
| tex->fMetadata.fAAType = static_cast<uint16_t>(GrAAType::kCoverage); |
| } |
| |
| for (GrOp* tmp = this->nextInChain(); tmp; tmp = tmp->nextInChain()) { |
| auto tex = static_cast<TextureOpImpl*>(tmp); |
| SkASSERT(tex->fMetadata.aaType() == GrAAType::kCoverage || |
| tex->fMetadata.aaType() == GrAAType::kNone); |
| tex->fMetadata.fAAType = static_cast<uint16_t>(GrAAType::kCoverage); |
| } |
| } |
| |
| CombineResult onCombineIfPossible(GrOp* t, SkArenaAlloc*, const GrCaps& caps) override { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| auto that = t->cast<TextureOpImpl>(); |
| |
| SkDEBUGCODE(this->validate();) |
| SkDEBUGCODE(that->validate();) |
| |
| if (fDesc || that->fDesc) { |
| // This should never happen (since only DDL recorded ops should be prePrepared) |
| // but, in any case, we should never combine ops that that been prePrepared |
| return CombineResult::kCannotCombine; |
| } |
| |
| if (fMetadata.subset() != that->fMetadata.subset()) { |
| // It is technically possible to combine operations across subset modes, but performance |
| // testing suggests it's better to make more draw calls where some take advantage of |
| // the more optimal shader path without coordinate clamping. |
| return CombineResult::kCannotCombine; |
| } |
| if (!GrColorSpaceXform::Equals(fTextureColorSpaceXform.get(), |
| that->fTextureColorSpaceXform.get())) { |
| return CombineResult::kCannotCombine; |
| } |
| |
| bool upgradeToCoverageAAOnMerge = false; |
| if (fMetadata.aaType() != that->fMetadata.aaType()) { |
| if (!CanUpgradeAAOnMerge(fMetadata.aaType(), that->fMetadata.aaType())) { |
| return CombineResult::kCannotCombine; |
| } |
| upgradeToCoverageAAOnMerge = true; |
| } |
| |
| if (CombinedQuadCountWillOverflow(fMetadata.aaType(), upgradeToCoverageAAOnMerge, |
| this->numChainedQuads() + that->numChainedQuads())) { |
| return CombineResult::kCannotCombine; |
| } |
| |
| if (fMetadata.saturate() != that->fMetadata.saturate()) { |
| return CombineResult::kCannotCombine; |
| } |
| if (fMetadata.filter() != that->fMetadata.filter()) { |
| return CombineResult::kCannotCombine; |
| } |
| if (fMetadata.mipmapMode() != that->fMetadata.mipmapMode()) { |
| return CombineResult::kCannotCombine; |
| } |
| if (fMetadata.fSwizzle != that->fMetadata.fSwizzle) { |
| return CombineResult::kCannotCombine; |
| } |
| const auto* thisProxy = fViewCountPairs[0].fProxy.get(); |
| const auto* thatProxy = that->fViewCountPairs[0].fProxy.get(); |
| if (fMetadata.fProxyCount > 1 || that->fMetadata.fProxyCount > 1 || |
| thisProxy != thatProxy) { |
| // We can't merge across different proxies. Check if 'this' can be chained with 'that'. |
| if (GrTextureProxy::ProxiesAreCompatibleAsDynamicState(thisProxy, thatProxy) && |
| caps.dynamicStateArrayGeometryProcessorTextureSupport() && |
| fMetadata.aaType() == that->fMetadata.aaType()) { |
| // We only allow chaining when the aaTypes match bc otherwise the AA type |
| // reported by the chain can be inconsistent. That is, since chaining doesn't |
| // propagate revised AA information throughout the chain, the head of the chain |
| // could have an AA setting of kNone while the chain as a whole could have a |
| // setting of kCoverage. This inconsistency would then interfere with the validity |
| // of the CombinedQuadCountWillOverflow calls. |
| // This problem doesn't occur w/ merging bc we do propagate the AA information |
| // (in propagateCoverageAAThroughoutChain) below. |
| return CombineResult::kMayChain; |
| } |
| return CombineResult::kCannotCombine; |
| } |
| |
| fMetadata.fSubset |= that->fMetadata.fSubset; |
| fMetadata.fColorType = std::max(fMetadata.fColorType, that->fMetadata.fColorType); |
| |
| // Concatenate quad lists together |
| fQuads.concat(that->fQuads); |
| fViewCountPairs[0].fQuadCnt += that->fQuads.count(); |
| fMetadata.fTotalQuadCount += that->fQuads.count(); |
| |
| if (upgradeToCoverageAAOnMerge) { |
| // This merger may be the start of a concatenation of two chains. When one |
| // of the chains mutates its AA the other must follow suit or else the above AA |
| // check may prevent later ops from chaining together. A specific example of this is |
| // when chain2 is prepended onto chain1: |
| // chain1 (that): opA (non-AA/mergeable) opB (non-AA/non-mergeable) |
| // chain2 (this): opC (cov-AA/non-mergeable) opD (cov-AA/mergeable) |
| // W/o this propagation, after opD & opA merge, opB and opC would say they couldn't |
| // chain - which would stop the concatenation process. |
| this->propagateCoverageAAThroughoutChain(); |
| that->propagateCoverageAAThroughoutChain(); |
| } |
| |
| SkDEBUGCODE(this->validate();) |
| |
| return CombineResult::kMerged; |
| } |
| |
| #if GR_TEST_UTILS |
| SkString onDumpInfo() const override { |
| SkString str = SkStringPrintf("# draws: %d\n", fQuads.count()); |
| auto iter = fQuads.iterator(); |
| for (unsigned p = 0; p < fMetadata.fProxyCount; ++p) { |
| SkString proxyStr = fViewCountPairs[p].fProxy->dump(); |
| str.append(proxyStr); |
| str.appendf(", Filter: %d, MM: %d\n", |
| static_cast<int>(fMetadata.fFilter), |
| static_cast<int>(fMetadata.fMipmapMode)); |
| for (int i = 0; i < fViewCountPairs[p].fQuadCnt && iter.next(); ++i) { |
| const GrQuad* quad = iter.deviceQuad(); |
| GrQuad uv = iter.isLocalValid() ? *(iter.localQuad()) : GrQuad(); |
| const ColorSubsetAndAA& info = iter.metadata(); |
| str.appendf( |
| "%d: Color: 0x%08x, Subset(%d): [L: %.2f, T: %.2f, R: %.2f, B: %.2f]\n" |
| " UVs [(%.2f, %.2f), (%.2f, %.2f), (%.2f, %.2f), (%.2f, %.2f)]\n" |
| " Quad [(%.2f, %.2f), (%.2f, %.2f), (%.2f, %.2f), (%.2f, %.2f)]\n", |
| i, info.fColor.toBytes_RGBA(), fMetadata.fSubset, info.fSubsetRect.fLeft, |
| info.fSubsetRect.fTop, info.fSubsetRect.fRight, info.fSubsetRect.fBottom, |
| quad->point(0).fX, quad->point(0).fY, quad->point(1).fX, quad->point(1).fY, |
| quad->point(2).fX, quad->point(2).fY, quad->point(3).fX, quad->point(3).fY, |
| uv.point(0).fX, uv.point(0).fY, uv.point(1).fX, uv.point(1).fY, |
| uv.point(2).fX, uv.point(2).fY, uv.point(3).fX, uv.point(3).fY); |
| } |
| } |
| return str; |
| } |
| #endif |
| |
| GrQuadBuffer<ColorSubsetAndAA> fQuads; |
| sk_sp<GrColorSpaceXform> fTextureColorSpaceXform; |
| // Most state of TextureOp is packed into these two field to minimize the op's size. |
| // Historically, increasing the size of TextureOp has caused surprising perf regressions, so |
| // consider/measure changes with care. |
| Desc* fDesc; |
| Metadata fMetadata; |
| |
| // This field must go last. When allocating this op, we will allocate extra space to hold |
| // additional ViewCountPairs immediately after the op's allocation so we can treat this |
| // as an fProxyCnt-length array. |
| ViewCountPair fViewCountPairs[1]; |
| |
| using INHERITED = GrMeshDrawOp; |
| }; |
| |
| } // anonymous namespace |
| |
| namespace skgpu::ganesh { |
| |
| #if GR_TEST_UTILS |
| uint32_t TextureOp::ClassID() { |
| return TextureOpImpl::ClassID(); |
| } |
| #endif |
| |
| GrOp::Owner TextureOp::Make(GrRecordingContext* context, |
| GrSurfaceProxyView proxyView, |
| SkAlphaType alphaType, |
| sk_sp<GrColorSpaceXform> textureXform, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| const SkPMColor4f& color, |
| Saturate saturate, |
| SkBlendMode blendMode, |
| GrAAType aaType, |
| DrawQuad* quad, |
| const SkRect* subset) { |
| // Apply optimizations that are valid whether or not using TextureOp or FillRectOp |
| if (subset && subset->contains(proxyView.proxy()->backingStoreBoundsRect())) { |
| // No need for a shader-based subset if hardware clamping achieves the same effect |
| subset = nullptr; |
| } |
| |
| if (filter != GrSamplerState::Filter::kNearest || mm != GrSamplerState::MipmapMode::kNone) { |
| auto [mustFilter, mustMM] = filter_and_mm_have_effect(quad->fLocal, quad->fDevice); |
| if (!mustFilter) { |
| filter = GrSamplerState::Filter::kNearest; |
| } |
| if (!mustMM) { |
| mm = GrSamplerState::MipmapMode::kNone; |
| } |
| } |
| |
| if (blendMode == SkBlendMode::kSrcOver) { |
| return TextureOpImpl::Make(context, std::move(proxyView), std::move(textureXform), filter, |
| mm, color, saturate, aaType, std::move(quad), subset); |
| } else { |
| // Emulate complex blending using FillRectOp |
| GrSamplerState samplerState(GrSamplerState::WrapMode::kClamp, filter, mm); |
| GrPaint paint; |
| paint.setColor4f(color); |
| paint.setXPFactory(SkBlendMode_AsXPFactory(blendMode)); |
| |
| std::unique_ptr<GrFragmentProcessor> fp; |
| const auto& caps = *context->priv().caps(); |
| if (subset) { |
| SkRect localRect; |
| if (quad->fLocal.asRect(&localRect)) { |
| fp = GrTextureEffect::MakeSubset(std::move(proxyView), alphaType, SkMatrix::I(), |
| samplerState, *subset, localRect, caps); |
| } else { |
| fp = GrTextureEffect::MakeSubset(std::move(proxyView), alphaType, SkMatrix::I(), |
| samplerState, *subset, caps); |
| } |
| } else { |
| fp = GrTextureEffect::Make(std::move(proxyView), alphaType, SkMatrix::I(), samplerState, |
| caps); |
| } |
| fp = GrColorSpaceXformEffect::Make(std::move(fp), std::move(textureXform)); |
| fp = GrBlendFragmentProcessor::Make<SkBlendMode::kModulate>(std::move(fp), nullptr); |
| if (saturate == Saturate::kYes) { |
| fp = GrFragmentProcessor::ClampOutput(std::move(fp)); |
| } |
| paint.setColorFragmentProcessor(std::move(fp)); |
| return v1::FillRectOp::Make(context, std::move(paint), aaType, quad); |
| } |
| } |
| |
| // A helper class that assists in breaking up bulk API quad draws into manageable chunks. |
| class TextureOp::BatchSizeLimiter { |
| public: |
| BatchSizeLimiter(v1::SurfaceDrawContext* sdc, |
| const GrClip* clip, |
| GrRecordingContext* rContext, |
| int numEntries, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| Saturate saturate, |
| SkCanvas::SrcRectConstraint constraint, |
| const SkMatrix& viewMatrix, |
| sk_sp<GrColorSpaceXform> textureColorSpaceXform) |
| : fSDC(sdc) |
| , fClip(clip) |
| , fContext(rContext) |
| , fFilter(filter) |
| , fMipmapMode(mm) |
| , fSaturate(saturate) |
| , fConstraint(constraint) |
| , fViewMatrix(viewMatrix) |
| , fTextureColorSpaceXform(textureColorSpaceXform) |
| , fNumLeft(numEntries) {} |
| |
| void createOp(GrTextureSetEntry set[], int clumpSize, GrAAType aaType) { |
| |
| int clumpProxyCount = proxy_run_count(&set[fNumClumped], clumpSize); |
| GrOp::Owner op = TextureOpImpl::Make(fContext, |
| &set[fNumClumped], |
| clumpSize, |
| clumpProxyCount, |
| fFilter, |
| fMipmapMode, |
| fSaturate, |
| aaType, |
| fConstraint, |
| fViewMatrix, |
| fTextureColorSpaceXform); |
| fSDC->addDrawOp(fClip, std::move(op)); |
| |
| fNumLeft -= clumpSize; |
| fNumClumped += clumpSize; |
| } |
| |
| int numLeft() const { return fNumLeft; } |
| int baseIndex() const { return fNumClumped; } |
| |
| private: |
| v1::SurfaceDrawContext* fSDC; |
| const GrClip* fClip; |
| GrRecordingContext* fContext; |
| GrSamplerState::Filter fFilter; |
| GrSamplerState::MipmapMode fMipmapMode; |
| Saturate fSaturate; |
| SkCanvas::SrcRectConstraint fConstraint; |
| const SkMatrix& fViewMatrix; |
| sk_sp<GrColorSpaceXform> fTextureColorSpaceXform; |
| |
| int fNumLeft; |
| int fNumClumped = 0; // also the offset for the start of the next clump |
| }; |
| |
| // Greedily clump quad draws together until the index buffer limit is exceeded. |
| void TextureOp::AddTextureSetOps(v1::SurfaceDrawContext* sdc, |
| const GrClip* clip, |
| GrRecordingContext* context, |
| GrTextureSetEntry set[], |
| int cnt, |
| int proxyRunCnt, |
| GrSamplerState::Filter filter, |
| GrSamplerState::MipmapMode mm, |
| Saturate saturate, |
| SkBlendMode blendMode, |
| GrAAType aaType, |
| SkCanvas::SrcRectConstraint constraint, |
| const SkMatrix& viewMatrix, |
| sk_sp<GrColorSpaceXform> textureColorSpaceXform) { |
| // Ensure that the index buffer limits are lower than the proxy and quad count limits of |
| // the op's metadata so we don't need to worry about overflow. |
| SkDEBUGCODE(TextureOpImpl::ValidateResourceLimits();) |
| SkASSERT(proxy_run_count(set, cnt) == proxyRunCnt); |
| |
| // First check if we can support batches as a single op |
| if (blendMode != SkBlendMode::kSrcOver || |
| !context->priv().caps()->dynamicStateArrayGeometryProcessorTextureSupport()) { |
| // Append each entry as its own op; these may still be GrTextureOps if the blend mode is |
| // src-over but the backend doesn't support dynamic state changes. Otherwise Make() |
| // automatically creates the appropriate FillRectOp to emulate TextureOp. |
| SkMatrix ctm; |
| for (int i = 0; i < cnt; ++i) { |
| ctm = viewMatrix; |
| if (set[i].fPreViewMatrix) { |
| ctm.preConcat(*set[i].fPreViewMatrix); |
| } |
| |
| DrawQuad quad; |
| quad.fEdgeFlags = set[i].fAAFlags; |
| if (set[i].fDstClipQuad) { |
| quad.fDevice = GrQuad::MakeFromSkQuad(set[i].fDstClipQuad, ctm); |
| |
| SkPoint srcPts[4]; |
| GrMapRectPoints(set[i].fDstRect, set[i].fSrcRect, set[i].fDstClipQuad, srcPts, 4); |
| quad.fLocal = GrQuad::MakeFromSkQuad(srcPts, SkMatrix::I()); |
| } else { |
| quad.fDevice = GrQuad::MakeFromRect(set[i].fDstRect, ctm); |
| quad.fLocal = GrQuad(set[i].fSrcRect); |
| } |
| |
| const SkRect* subset = constraint == SkCanvas::kStrict_SrcRectConstraint |
| ? &set[i].fSrcRect : nullptr; |
| |
| auto op = Make(context, set[i].fProxyView, set[i].fSrcAlphaType, textureColorSpaceXform, |
| filter, mm, set[i].fColor, saturate, blendMode, aaType, &quad, subset); |
| sdc->addDrawOp(clip, std::move(op)); |
| } |
| return; |
| } |
| |
| // Second check if we can always just make a single op and avoid the extra iteration |
| // needed to clump things together. |
| if (cnt <= std::min(GrResourceProvider::MaxNumNonAAQuads(), |
| GrResourceProvider::MaxNumAAQuads())) { |
| auto op = TextureOpImpl::Make(context, set, cnt, proxyRunCnt, filter, mm, saturate, aaType, |
| constraint, viewMatrix, std::move(textureColorSpaceXform)); |
| sdc->addDrawOp(clip, std::move(op)); |
| return; |
| } |
| |
| BatchSizeLimiter state(sdc, clip, context, cnt, filter, mm, saturate, constraint, viewMatrix, |
| std::move(textureColorSpaceXform)); |
| |
| // kNone and kMSAA never get altered |
| if (aaType == GrAAType::kNone || aaType == GrAAType::kMSAA) { |
| // Clump these into series of MaxNumNonAAQuads-sized GrTextureOps |
| while (state.numLeft() > 0) { |
| int clumpSize = std::min(state.numLeft(), GrResourceProvider::MaxNumNonAAQuads()); |
| |
| state.createOp(set, clumpSize, aaType); |
| } |
| } else { |
| // kCoverage can be downgraded to kNone. Note that the following is conservative. kCoverage |
| // can also get downgraded to kNone if all the quads are on integer coordinates and |
| // axis-aligned. |
| SkASSERT(aaType == GrAAType::kCoverage); |
| |
| while (state.numLeft() > 0) { |
| GrAAType runningAA = GrAAType::kNone; |
| bool clumped = false; |
| |
| for (int i = 0; i < state.numLeft(); ++i) { |
| int absIndex = state.baseIndex() + i; |
| |
| if (set[absIndex].fAAFlags != GrQuadAAFlags::kNone || |
| runningAA == GrAAType::kCoverage) { |
| |
| if (i >= GrResourceProvider::MaxNumAAQuads()) { |
| // Here we either need to boost the AA type to kCoverage, but doing so with |
| // all the accumulated quads would overflow, or we have a set of AA quads |
| // that has just gotten too large. In either case, calve off the existing |
| // quads as their own TextureOp. |
| state.createOp( |
| set, |
| runningAA == GrAAType::kNone ? i : GrResourceProvider::MaxNumAAQuads(), |
| runningAA); // maybe downgrading AA here |
| clumped = true; |
| break; |
| } |
| |
| runningAA = GrAAType::kCoverage; |
| } else if (runningAA == GrAAType::kNone) { |
| |
| if (i >= GrResourceProvider::MaxNumNonAAQuads()) { |
| // Here we've found a consistent batch of non-AA quads that has gotten too |
| // large. Calve it off as its own TextureOp. |
| state.createOp(set, GrResourceProvider::MaxNumNonAAQuads(), |
| GrAAType::kNone); // definitely downgrading AA here |
| clumped = true; |
| break; |
| } |
| } |
| } |
| |
| if (!clumped) { |
| // We ran through the above loop w/o hitting a limit. Spit out this last clump of |
| // quads and call it a day. |
| state.createOp(set, state.numLeft(), runningAA); // maybe downgrading AA here |
| } |
| } |
| } |
| } |
| |
| } // namespace skgpu::ganesh |
| |
| #if GR_TEST_UTILS |
| GR_DRAW_OP_TEST_DEFINE(TextureOpImpl) { |
| SkISize dims; |
| dims.fHeight = random->nextULessThan(90) + 10; |
| dims.fWidth = random->nextULessThan(90) + 10; |
| auto origin = random->nextBool() ? kTopLeft_GrSurfaceOrigin : kBottomLeft_GrSurfaceOrigin; |
| GrMipmapped mipmapped = random->nextBool() ? GrMipmapped::kYes : GrMipmapped::kNo; |
| SkBackingFit fit = SkBackingFit::kExact; |
| if (mipmapped == GrMipmapped::kNo) { |
| fit = random->nextBool() ? SkBackingFit::kApprox : SkBackingFit::kExact; |
| } |
| const GrBackendFormat format = |
| context->priv().caps()->getDefaultBackendFormat(GrColorType::kRGBA_8888, |
| GrRenderable::kNo); |
| GrProxyProvider* proxyProvider = context->priv().proxyProvider(); |
| sk_sp<GrTextureProxy> proxy = proxyProvider->createProxy(format, |
| dims, |
| GrRenderable::kNo, |
| 1, |
| mipmapped, |
| fit, |
| skgpu::Budgeted::kNo, |
| GrProtected::kNo, |
| /*label=*/"TextureOp", |
| GrInternalSurfaceFlags::kNone); |
| |
| SkRect rect = GrTest::TestRect(random); |
| SkRect srcRect; |
| srcRect.fLeft = random->nextRangeScalar(0.f, proxy->width() / 2.f); |
| srcRect.fRight = random->nextRangeScalar(0.f, proxy->width()) + proxy->width() / 2.f; |
| srcRect.fTop = random->nextRangeScalar(0.f, proxy->height() / 2.f); |
| srcRect.fBottom = random->nextRangeScalar(0.f, proxy->height()) + proxy->height() / 2.f; |
| SkMatrix viewMatrix = GrTest::TestMatrixPreservesRightAngles(random); |
| SkPMColor4f color = SkPMColor4f::FromBytes_RGBA(SkColorToPremulGrColor(random->nextU())); |
| GrSamplerState::Filter filter = (GrSamplerState::Filter)random->nextULessThan( |
| static_cast<uint32_t>(GrSamplerState::Filter::kLast) + 1); |
| GrSamplerState::MipmapMode mm = GrSamplerState::MipmapMode::kNone; |
| if (mipmapped == GrMipmapped::kYes) { |
| mm = (GrSamplerState::MipmapMode)random->nextULessThan( |
| static_cast<uint32_t>(GrSamplerState::MipmapMode::kLast) + 1); |
| } |
| |
| auto texXform = GrTest::TestColorXform(random); |
| GrAAType aaType = GrAAType::kNone; |
| if (random->nextBool()) { |
| aaType = (numSamples > 1) ? GrAAType::kMSAA : GrAAType::kCoverage; |
| } |
| GrQuadAAFlags aaFlags = GrQuadAAFlags::kNone; |
| aaFlags |= random->nextBool() ? GrQuadAAFlags::kLeft : GrQuadAAFlags::kNone; |
| aaFlags |= random->nextBool() ? GrQuadAAFlags::kTop : GrQuadAAFlags::kNone; |
| aaFlags |= random->nextBool() ? GrQuadAAFlags::kRight : GrQuadAAFlags::kNone; |
| aaFlags |= random->nextBool() ? GrQuadAAFlags::kBottom : GrQuadAAFlags::kNone; |
| bool useSubset = random->nextBool(); |
| auto saturate = random->nextBool() ? TextureOp::Saturate::kYes |
| : TextureOp::Saturate::kNo; |
| GrSurfaceProxyView proxyView( |
| std::move(proxy), origin, |
| context->priv().caps()->getReadSwizzle(format, GrColorType::kRGBA_8888)); |
| auto alphaType = static_cast<SkAlphaType>( |
| random->nextRangeU(kUnknown_SkAlphaType + 1, kLastEnum_SkAlphaType)); |
| |
| DrawQuad quad = {GrQuad::MakeFromRect(rect, viewMatrix), GrQuad(srcRect), aaFlags}; |
| return TextureOp::Make(context, std::move(proxyView), alphaType, |
| std::move(texXform), filter, mm, color, saturate, |
| SkBlendMode::kSrcOver, aaType, &quad, |
| useSubset ? &srcRect : nullptr); |
| } |
| |
| #endif // GR_TEST_UTILS |