|  | /* | 
|  | * Copyright 2018 Google Inc. | 
|  | * | 
|  | * Use of this source code is governed by a BSD-style license that can be | 
|  | * found in the LICENSE file. | 
|  | */ | 
|  |  | 
|  | #include "tools/DDLPromiseImageHelper.h" | 
|  |  | 
|  | #include "include/core/SkDeferredDisplayListRecorder.h" | 
|  | #include "include/core/SkPicture.h" | 
|  | #include "include/core/SkSerialProcs.h" | 
|  | #include "include/core/SkYUVAIndex.h" | 
|  | #include "include/core/SkYUVASizeInfo.h" | 
|  | #include "include/gpu/GrDirectContext.h" | 
|  | #include "src/codec/SkCodecImageGenerator.h" | 
|  | #include "src/core/SkCachedData.h" | 
|  | #include "src/core/SkMipmap.h" | 
|  | #include "src/core/SkTaskGroup.h" | 
|  | #include "src/gpu/GrContextPriv.h" | 
|  | #include "src/image/SkImage_Base.h" | 
|  | #include "src/image/SkImage_GpuYUVA.h" | 
|  |  | 
|  | DDLPromiseImageHelper::PromiseImageInfo::PromiseImageInfo(int index, | 
|  | uint32_t originalUniqueID, | 
|  | const SkImageInfo& ii) | 
|  | : fIndex(index) | 
|  | , fOriginalUniqueID(originalUniqueID) | 
|  | , fImageInfo(ii) { | 
|  | } | 
|  |  | 
|  | DDLPromiseImageHelper::PromiseImageInfo::PromiseImageInfo(PromiseImageInfo&& other) | 
|  | : fIndex(other.fIndex) | 
|  | , fOriginalUniqueID(other.fOriginalUniqueID) | 
|  | , fImageInfo(other.fImageInfo) | 
|  | , fBaseLevel(other.fBaseLevel) | 
|  | , fMipLevels(std::move(other.fMipLevels)) | 
|  | , fYUVAPixmaps(std::move(other.fYUVAPixmaps)) { | 
|  | for (int i = 0; i < SkYUVASizeInfo::kMaxCount; ++i) { | 
|  | fCallbackContexts[i] = std::move(other.fCallbackContexts[i]); | 
|  | } | 
|  | std::copy_n(other.fYUVAIndices, SkYUVAIndex::kIndexCount, fYUVAIndices); | 
|  | } | 
|  |  | 
|  | DDLPromiseImageHelper::PromiseImageInfo::~PromiseImageInfo() {} | 
|  |  | 
|  | std::unique_ptr<SkPixmap[]> DDLPromiseImageHelper::PromiseImageInfo::normalMipLevels() const { | 
|  | SkASSERT(!this->isYUV()); | 
|  | std::unique_ptr<SkPixmap[]> pixmaps(new SkPixmap[this->numMipLevels()]); | 
|  | pixmaps[0] = fBaseLevel.pixmap(); | 
|  | if (fMipLevels) { | 
|  | for (int i = 0; i < fMipLevels->countLevels(); ++i) { | 
|  | SkMipmap::Level mipLevel; | 
|  | fMipLevels->getLevel(i, &mipLevel); | 
|  | pixmaps[i+1] = mipLevel.fPixmap; | 
|  | } | 
|  | } | 
|  | return pixmaps; | 
|  | } | 
|  |  | 
|  | int DDLPromiseImageHelper::PromiseImageInfo::numMipLevels() const { | 
|  | SkASSERT(!this->isYUV()); | 
|  | return fMipLevels ? fMipLevels->countLevels()+1 : 1; | 
|  | } | 
|  |  | 
|  | void DDLPromiseImageHelper::PromiseImageInfo::setMipLevels(const SkBitmap& baseLevel, | 
|  | std::unique_ptr<SkMipmap> mipLevels) { | 
|  | fBaseLevel = baseLevel; | 
|  | fMipLevels = std::move(mipLevels); | 
|  | } | 
|  |  | 
|  | void DDLPromiseImageHelper::PromiseImageInfo::initYUVAIndices() { | 
|  | SkYUVASizeInfo unusedSizeInfo; | 
|  | fYUVAPixmaps.toLegacy(&unusedSizeInfo, fYUVAIndices); | 
|  | // We can wind up with alpha pixmaps that become red textures. If so, update YUVA indices. | 
|  | // Note: This goes away when we start creating YUVA images using SkYUVAInfo. | 
|  | for (int i = 0; i < SkYUVAIndex::kIndexCount; ++i) { | 
|  | if (fYUVAIndices[i].fIndex >= 0) { | 
|  | int textureIdx = fYUVAIndices[i].fIndex; | 
|  | const auto& format = fCallbackContexts[textureIdx]->backendFormat(); | 
|  | if (fYUVAIndices[i].fChannel == SkColorChannel::kA && | 
|  | format.channelMask() == SkColorChannelFlag::kRed_SkColorChannelFlag) { | 
|  | fYUVAIndices[i].fChannel = SkColorChannel::kR; | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | /////////////////////////////////////////////////////////////////////////////////////////////////// | 
|  | PromiseImageCallbackContext::~PromiseImageCallbackContext() { | 
|  | SkASSERT(fDoneCnt == fNumImages); | 
|  | SkASSERT(!fUnreleasedFulfills); | 
|  | SkASSERT(fTotalReleases == fTotalFulfills); | 
|  | SkASSERT(!fTotalFulfills || fDoneCnt); | 
|  |  | 
|  | if (fPromiseImageTexture) { | 
|  | fContext->deleteBackendTexture(fPromiseImageTexture->backendTexture()); | 
|  | } | 
|  | } | 
|  |  | 
|  | void PromiseImageCallbackContext::setBackendTexture(const GrBackendTexture& backendTexture) { | 
|  | SkASSERT(!fPromiseImageTexture); | 
|  | SkASSERT(fBackendFormat == backendTexture.getBackendFormat()); | 
|  | fPromiseImageTexture = SkPromiseImageTexture::Make(backendTexture); | 
|  | } | 
|  |  | 
|  | void PromiseImageCallbackContext::destroyBackendTexture() { | 
|  | SkASSERT(!fPromiseImageTexture || fPromiseImageTexture->unique()); | 
|  |  | 
|  | if (fPromiseImageTexture) { | 
|  | fContext->deleteBackendTexture(fPromiseImageTexture->backendTexture()); | 
|  | } | 
|  | fPromiseImageTexture = nullptr; | 
|  | } | 
|  |  | 
|  | /////////////////////////////////////////////////////////////////////////////////////////////////// | 
|  |  | 
|  | sk_sp<SkData> DDLPromiseImageHelper::deflateSKP(const SkPicture* inputPicture) { | 
|  | SkSerialProcs procs; | 
|  |  | 
|  | procs.fImageCtx = this; | 
|  | procs.fImageProc = [](SkImage* image, void* ctx) -> sk_sp<SkData> { | 
|  | auto helper = static_cast<DDLPromiseImageHelper*>(ctx); | 
|  |  | 
|  | int id = helper->findOrDefineImage(image); | 
|  |  | 
|  | // Even if 'id' is invalid (i.e., -1) write it to the SKP | 
|  | return SkData::MakeWithCopy(&id, sizeof(id)); | 
|  | }; | 
|  |  | 
|  | return inputPicture->serialize(&procs); | 
|  | } | 
|  |  | 
|  | static GrBackendTexture create_yuva_texture(GrDirectContext* direct, const SkPixmap& pm, | 
|  | const SkYUVAIndex yuvaIndices[4], int texIndex) { | 
|  | SkASSERT(texIndex >= 0 && texIndex <= 3); | 
|  |  | 
|  | #ifdef SK_DEBUG | 
|  | int channelCount = 0; | 
|  | for (int i = 0; i < SkYUVAIndex::kIndexCount; ++i) { | 
|  | if (yuvaIndices[i].fIndex == texIndex) { | 
|  | ++channelCount; | 
|  | } | 
|  | } | 
|  | if (2 == channelCount) { | 
|  | SkASSERT(kR8G8_unorm_SkColorType == pm.colorType()); | 
|  | } | 
|  | #endif | 
|  | bool finishedBECreate = false; | 
|  | auto markFinished = [](void* context) { | 
|  | *(bool*)context = true; | 
|  | }; | 
|  | auto beTex = direct->createBackendTexture(&pm, 1, GrRenderable::kNo, GrProtected::kNo, | 
|  | markFinished, &finishedBECreate); | 
|  | if (beTex.isValid()) { | 
|  | direct->submit(); | 
|  | while (!finishedBECreate) { | 
|  | direct->checkAsyncWorkCompletion(); | 
|  | } | 
|  | } | 
|  | return beTex; | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Create backend textures and upload data to them for all the textures required to satisfy | 
|  | * a single promise image. | 
|  | * For YUV textures this will result in up to 4 actual textures. | 
|  | */ | 
|  | void DDLPromiseImageHelper::CreateBETexturesForPromiseImage(GrDirectContext* direct, | 
|  | PromiseImageInfo* info) { | 
|  | if (info->isYUV()) { | 
|  | int numPixmaps; | 
|  | SkAssertResult(SkYUVAIndex::AreValidIndices(info->yuvaIndices(), &numPixmaps)); | 
|  | for (int j = 0; j < numPixmaps; ++j) { | 
|  | const SkPixmap& yuvPixmap = info->yuvPixmap(j); | 
|  |  | 
|  | PromiseImageCallbackContext* callbackContext = info->callbackContext(j); | 
|  | SkASSERT(callbackContext); | 
|  |  | 
|  | // DDL TODO: what should we do with mipmapped YUV images | 
|  | callbackContext->setBackendTexture(create_yuva_texture(direct, yuvPixmap, | 
|  | info->yuvaIndices(), j)); | 
|  | SkASSERT(callbackContext->promiseImageTexture()); | 
|  | } | 
|  | } else { | 
|  | PromiseImageCallbackContext* callbackContext = info->callbackContext(0); | 
|  | if (!callbackContext) { | 
|  | // This texture would've been too large to fit on the GPU | 
|  | return; | 
|  | } | 
|  |  | 
|  | std::unique_ptr<SkPixmap[]> mipLevels = info->normalMipLevels(); | 
|  |  | 
|  | bool finishedBECreate = false; | 
|  | auto markFinished = [](void* context) { | 
|  | *(bool*)context = true; | 
|  | }; | 
|  | auto backendTex = direct->createBackendTexture(mipLevels.get(), info->numMipLevels(), | 
|  | GrRenderable::kNo, GrProtected::kNo, | 
|  | markFinished, &finishedBECreate); | 
|  | SkASSERT(backendTex.isValid()); | 
|  | direct->submit(); | 
|  | while (!finishedBECreate) { | 
|  | direct->checkAsyncWorkCompletion(); | 
|  | } | 
|  |  | 
|  | callbackContext->setBackendTexture(backendTex); | 
|  | } | 
|  | } | 
|  |  | 
|  | void DDLPromiseImageHelper::DeleteBETexturesForPromiseImage(GrDirectContext* direct, | 
|  | PromiseImageInfo* info) { | 
|  | if (info->isYUV()) { | 
|  | int numPixmaps; | 
|  | SkAssertResult(SkYUVAIndex::AreValidIndices(info->yuvaIndices(), &numPixmaps)); | 
|  | for (int j = 0; j < numPixmaps; ++j) { | 
|  | PromiseImageCallbackContext* callbackContext = info->callbackContext(j); | 
|  | SkASSERT(callbackContext); | 
|  |  | 
|  | callbackContext->destroyBackendTexture(); | 
|  | SkASSERT(!callbackContext->promiseImageTexture()); | 
|  | } | 
|  | } else { | 
|  | PromiseImageCallbackContext* callbackContext = info->callbackContext(0); | 
|  | if (!callbackContext) { | 
|  | // This texture would've been too large to fit on the GPU | 
|  | return; | 
|  | } | 
|  |  | 
|  | callbackContext->destroyBackendTexture(); | 
|  | SkASSERT(!callbackContext->promiseImageTexture()); | 
|  | } | 
|  | } | 
|  |  | 
|  | void DDLPromiseImageHelper::createCallbackContexts(GrDirectContext* direct) { | 
|  | const GrCaps* caps = direct->priv().caps(); | 
|  | const int maxDimension = caps->maxTextureSize(); | 
|  |  | 
|  | for (int i = 0; i < fImageInfo.count(); ++i) { | 
|  | PromiseImageInfo& info = fImageInfo[i]; | 
|  |  | 
|  | if (info.isYUV()) { | 
|  | int numPixmaps = info.numYUVAPlanes(); | 
|  |  | 
|  | for (int j = 0; j < numPixmaps; ++j) { | 
|  | const SkPixmap& yuvPixmap = info.yuvPixmap(j); | 
|  |  | 
|  | GrBackendFormat backendFormat = direct->defaultBackendFormat(yuvPixmap.colorType(), | 
|  | GrRenderable::kNo); | 
|  |  | 
|  | sk_sp<PromiseImageCallbackContext> callbackContext( | 
|  | new PromiseImageCallbackContext(direct, backendFormat)); | 
|  |  | 
|  | info.setCallbackContext(j, std::move(callbackContext)); | 
|  | } | 
|  | info.initYUVAIndices(); | 
|  | } else { | 
|  | const SkBitmap& baseLevel = info.baseLevel(); | 
|  |  | 
|  | // TODO: explicitly mark the PromiseImageInfo as too big and check in uploadAllToGPU | 
|  | if (maxDimension < std::max(baseLevel.width(), baseLevel.height())) { | 
|  | // This won't fit on the GPU. Fallback to a raster-backed image per tile. | 
|  | continue; | 
|  | } | 
|  |  | 
|  | GrBackendFormat backendFormat = direct->defaultBackendFormat(baseLevel.colorType(), | 
|  | GrRenderable::kNo); | 
|  | if (!caps->isFormatTexturable(backendFormat)) { | 
|  | continue; | 
|  | } | 
|  |  | 
|  | sk_sp<PromiseImageCallbackContext> callbackContext( | 
|  | new PromiseImageCallbackContext(direct, backendFormat)); | 
|  |  | 
|  | info.setCallbackContext(0, std::move(callbackContext)); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void DDLPromiseImageHelper::uploadAllToGPU(SkTaskGroup* taskGroup, GrDirectContext* direct) { | 
|  | if (taskGroup) { | 
|  | for (int i = 0; i < fImageInfo.count(); ++i) { | 
|  | PromiseImageInfo* info = &fImageInfo[i]; | 
|  |  | 
|  | taskGroup->add([direct, info]() { CreateBETexturesForPromiseImage(direct, info); }); | 
|  | } | 
|  | } else { | 
|  | for (int i = 0; i < fImageInfo.count(); ++i) { | 
|  | CreateBETexturesForPromiseImage(direct, &fImageInfo[i]); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void DDLPromiseImageHelper::deleteAllFromGPU(SkTaskGroup* taskGroup, GrDirectContext* direct) { | 
|  | if (taskGroup) { | 
|  | for (int i = 0; i < fImageInfo.count(); ++i) { | 
|  | PromiseImageInfo* info = &fImageInfo[i]; | 
|  |  | 
|  | taskGroup->add([direct, info]() { DeleteBETexturesForPromiseImage(direct, info); }); | 
|  | } | 
|  | } else { | 
|  | for (int i = 0; i < fImageInfo.count(); ++i) { | 
|  | DeleteBETexturesForPromiseImage(direct, &fImageInfo[i]); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | sk_sp<SkPicture> DDLPromiseImageHelper::reinflateSKP( | 
|  | SkDeferredDisplayListRecorder* recorder, | 
|  | SkData* compressedPictureData, | 
|  | SkTArray<sk_sp<SkImage>>* promiseImages) const { | 
|  | PerRecorderContext perRecorderContext { recorder, this, promiseImages }; | 
|  |  | 
|  | SkDeserialProcs procs; | 
|  | procs.fImageCtx = (void*) &perRecorderContext; | 
|  | procs.fImageProc = CreatePromiseImages; | 
|  |  | 
|  | return SkPicture::MakeFromData(compressedPictureData, &procs); | 
|  | } | 
|  |  | 
|  | // This generates promise images to replace the indices in the compressed picture. This | 
|  | // reconstitution is performed separately in each thread so we end up with multiple | 
|  | // promise images referring to the same GrBackendTexture. | 
|  | sk_sp<SkImage> DDLPromiseImageHelper::CreatePromiseImages(const void* rawData, | 
|  | size_t length, void* ctxIn) { | 
|  | PerRecorderContext* perRecorderContext = static_cast<PerRecorderContext*>(ctxIn); | 
|  | const DDLPromiseImageHelper* helper = perRecorderContext->fHelper; | 
|  | SkDeferredDisplayListRecorder* recorder = perRecorderContext->fRecorder; | 
|  |  | 
|  | SkASSERT(length == sizeof(int)); | 
|  |  | 
|  | const int* indexPtr = static_cast<const int*>(rawData); | 
|  | if (!helper->isValidID(*indexPtr)) { | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | const DDLPromiseImageHelper::PromiseImageInfo& curImage = helper->getInfo(*indexPtr); | 
|  |  | 
|  | // If there is no callback context that means 'createCallbackContexts' determined the | 
|  | // texture wouldn't fit on the GPU. Create a separate bitmap-backed image for each thread. | 
|  | if (!curImage.isYUV() && !curImage.callbackContext(0)) { | 
|  | SkASSERT(curImage.baseLevel().isImmutable()); | 
|  | return SkImage::MakeFromBitmap(curImage.baseLevel()); | 
|  | } | 
|  |  | 
|  | SkASSERT(curImage.index() == *indexPtr); | 
|  |  | 
|  | sk_sp<SkImage> image; | 
|  | if (curImage.isYUV()) { | 
|  | GrBackendFormat backendFormats[SkYUVASizeInfo::kMaxCount]; | 
|  | void* contexts[SkYUVASizeInfo::kMaxCount] = { nullptr, nullptr, nullptr, nullptr }; | 
|  | SkISize sizes[SkYUVASizeInfo::kMaxCount]; | 
|  | // TODO: store this value somewhere? | 
|  | int textureCount; | 
|  | SkAssertResult(SkYUVAIndex::AreValidIndices(curImage.yuvaIndices(), &textureCount)); | 
|  | for (int i = 0; i < textureCount; ++i) { | 
|  | backendFormats[i] = curImage.backendFormat(i); | 
|  | SkASSERT(backendFormats[i].isValid()); | 
|  | contexts[i] = curImage.refCallbackContext(i).release(); | 
|  | sizes[i].set(curImage.yuvPixmap(i).width(), curImage.yuvPixmap(i).height()); | 
|  | } | 
|  | for (int i = textureCount; i < SkYUVASizeInfo::kMaxCount; ++i) { | 
|  | sizes[i] = SkISize::MakeEmpty(); | 
|  | } | 
|  |  | 
|  | image = recorder->makeYUVAPromiseTexture( | 
|  | curImage.yuvColorSpace(), | 
|  | backendFormats, | 
|  | sizes, | 
|  | curImage.yuvaIndices(), | 
|  | curImage.overallWidth(), | 
|  | curImage.overallHeight(), | 
|  | GrSurfaceOrigin::kTopLeft_GrSurfaceOrigin, | 
|  | curImage.refOverallColorSpace(), | 
|  | PromiseImageCallbackContext::PromiseImageFulfillProc, | 
|  | PromiseImageCallbackContext::PromiseImageReleaseProc, | 
|  | PromiseImageCallbackContext::PromiseImageDoneProc, | 
|  | contexts, | 
|  | SkDeferredDisplayListRecorder::PromiseImageApiVersion::kNew); | 
|  | for (int i = 0; i < textureCount; ++i) { | 
|  | curImage.callbackContext(i)->wasAddedToImage(); | 
|  | } | 
|  |  | 
|  | #ifdef SK_DEBUG | 
|  | { | 
|  | // By the peekProxy contract this image should not have a single backing proxy so | 
|  | // should return null. The call should also not trigger the conversion to RGBA. | 
|  | SkImage_GpuYUVA* yuva = reinterpret_cast<SkImage_GpuYUVA*>(image.get()); | 
|  | SkASSERT(!yuva->peekProxy()); | 
|  | SkASSERT(!yuva->peekProxy());  // the first call didn't force a conversion to RGBA | 
|  | } | 
|  | #endif | 
|  | } else { | 
|  | const GrBackendFormat& backendFormat = curImage.backendFormat(0); | 
|  | SkASSERT(backendFormat.isValid()); | 
|  |  | 
|  | // Each DDL recorder gets its own ref on the promise callback context for the | 
|  | // promise images it creates. | 
|  | image = recorder->makePromiseTexture( | 
|  | backendFormat, | 
|  | curImage.overallWidth(), | 
|  | curImage.overallHeight(), | 
|  | curImage.mipMapped(0), | 
|  | GrSurfaceOrigin::kTopLeft_GrSurfaceOrigin, | 
|  | curImage.overallColorType(), | 
|  | curImage.overallAlphaType(), | 
|  | curImage.refOverallColorSpace(), | 
|  | PromiseImageCallbackContext::PromiseImageFulfillProc, | 
|  | PromiseImageCallbackContext::PromiseImageReleaseProc, | 
|  | PromiseImageCallbackContext::PromiseImageDoneProc, | 
|  | (void*)curImage.refCallbackContext(0).release(), | 
|  | SkDeferredDisplayListRecorder::PromiseImageApiVersion::kNew); | 
|  | curImage.callbackContext(0)->wasAddedToImage(); | 
|  | } | 
|  | perRecorderContext->fPromiseImages->push_back(image); | 
|  | SkASSERT(image); | 
|  | return image; | 
|  | } | 
|  |  | 
|  | int DDLPromiseImageHelper::findImage(SkImage* image) const { | 
|  | for (int i = 0; i < fImageInfo.count(); ++i) { | 
|  | if (fImageInfo[i].originalUniqueID() == image->uniqueID()) { // trying to dedup here | 
|  | SkASSERT(fImageInfo[i].index() == i); | 
|  | SkASSERT(this->isValidID(i) && this->isValidID(fImageInfo[i].index())); | 
|  | return i; | 
|  | } | 
|  | } | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | int DDLPromiseImageHelper::addImage(SkImage* image) { | 
|  | SkImage_Base* ib = as_IB(image); | 
|  |  | 
|  | SkImageInfo overallII = SkImageInfo::Make(image->width(), image->height(), | 
|  | image->colorType() == kBGRA_8888_SkColorType | 
|  | ? kRGBA_8888_SkColorType | 
|  | : image->colorType(), | 
|  | image->alphaType(), | 
|  | image->refColorSpace()); | 
|  |  | 
|  | PromiseImageInfo& newImageInfo = fImageInfo.emplace_back(fImageInfo.count(), | 
|  | image->uniqueID(), | 
|  | overallII); | 
|  |  | 
|  | auto codec = SkCodecImageGenerator::MakeFromEncodedCodec(ib->refEncodedData()); | 
|  | SkYUVAPixmapInfo yuvaInfo; | 
|  | if (codec && codec->queryYUVAInfo(fSupportedYUVADataTypes, &yuvaInfo)) { | 
|  | auto yuvaPixmaps = SkYUVAPixmaps::Allocate(yuvaInfo); | 
|  | SkAssertResult(codec->getYUVAPlanes(yuvaPixmaps)); | 
|  | SkASSERT(yuvaPixmaps.isValid()); | 
|  | newImageInfo.setYUVPlanes(std::move(yuvaPixmaps)); | 
|  | } else { | 
|  | sk_sp<SkImage> rasterImage = image->makeRasterImage(); // force decoding of lazy images | 
|  | if (!rasterImage) { | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | SkBitmap tmp; | 
|  | tmp.allocPixels(overallII); | 
|  |  | 
|  | if (!rasterImage->readPixels(nullptr, tmp.pixmap(), 0, 0)) { | 
|  | return -1; | 
|  | } | 
|  |  | 
|  | tmp.setImmutable(); | 
|  |  | 
|  | // Given how the DDL testing harness works (i.e., only modifying the SkImages w/in an | 
|  | // SKP) we don't know if a given SkImage will require mipmapping. To work around this | 
|  | // we just try to create all the backend textures as mipmapped but, failing that, fall | 
|  | // back to un-mipped. | 
|  | std::unique_ptr<SkMipmap> mipmaps(SkMipmap::Build(tmp.pixmap(), nullptr)); | 
|  |  | 
|  | newImageInfo.setMipLevels(tmp, std::move(mipmaps)); | 
|  | } | 
|  | // In either case newImageInfo's PromiseImageCallbackContext is filled in by uploadAllToGPU | 
|  |  | 
|  | return fImageInfo.count()-1; | 
|  | } | 
|  |  | 
|  | int DDLPromiseImageHelper::findOrDefineImage(SkImage* image) { | 
|  | int preExistingID = this->findImage(image); | 
|  | if (preExistingID >= 0) { | 
|  | SkASSERT(this->isValidID(preExistingID)); | 
|  | return preExistingID; | 
|  | } | 
|  |  | 
|  | int newID = this->addImage(image); | 
|  | return newID; | 
|  | } |