You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1237 lines
47 KiB
1237 lines
47 KiB
/*
|
|
* Copyright (C) 2013 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include "rsdAllocation.h"
|
|
#include "rsdCore.h"
|
|
|
|
#include <android/native_window.h>
|
|
|
|
#ifdef RS_COMPATIBILITY_LIB
|
|
#include "rsCompatibilityLib.h"
|
|
#else
|
|
#include "rsdFrameBufferObj.h"
|
|
#include <vndk/window.h>
|
|
|
|
#include <GLES/gl.h>
|
|
#include <GLES2/gl2.h>
|
|
#include <GLES/glext.h>
|
|
#endif
|
|
|
|
#include <malloc.h> // for memalign()
|
|
#include <unistd.h> // for close()
|
|
|
|
using android::renderscript::Allocation;
|
|
using android::renderscript::Context;
|
|
using android::renderscript::Element;
|
|
using android::renderscript::Type;
|
|
using android::renderscript::rs_allocation;
|
|
using android::renderscript::rsBoxFilter565;
|
|
using android::renderscript::rsBoxFilter8888;
|
|
using android::renderscript::rsMax;
|
|
using android::renderscript::rsRound;
|
|
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
const static GLenum gFaceOrder[] = {
|
|
GL_TEXTURE_CUBE_MAP_POSITIVE_X,
|
|
GL_TEXTURE_CUBE_MAP_NEGATIVE_X,
|
|
GL_TEXTURE_CUBE_MAP_POSITIVE_Y,
|
|
GL_TEXTURE_CUBE_MAP_NEGATIVE_Y,
|
|
GL_TEXTURE_CUBE_MAP_POSITIVE_Z,
|
|
GL_TEXTURE_CUBE_MAP_NEGATIVE_Z
|
|
};
|
|
|
|
GLenum rsdTypeToGLType(RsDataType t) {
|
|
switch (t) {
|
|
case RS_TYPE_UNSIGNED_5_6_5: return GL_UNSIGNED_SHORT_5_6_5;
|
|
case RS_TYPE_UNSIGNED_5_5_5_1: return GL_UNSIGNED_SHORT_5_5_5_1;
|
|
case RS_TYPE_UNSIGNED_4_4_4_4: return GL_UNSIGNED_SHORT_4_4_4_4;
|
|
|
|
//case RS_TYPE_FLOAT_16: return GL_HALF_FLOAT;
|
|
case RS_TYPE_FLOAT_32: return GL_FLOAT;
|
|
case RS_TYPE_UNSIGNED_8: return GL_UNSIGNED_BYTE;
|
|
case RS_TYPE_UNSIGNED_16: return GL_UNSIGNED_SHORT;
|
|
case RS_TYPE_SIGNED_8: return GL_BYTE;
|
|
case RS_TYPE_SIGNED_16: return GL_SHORT;
|
|
default: break;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
GLenum rsdKindToGLFormat(RsDataKind k) {
|
|
switch (k) {
|
|
case RS_KIND_PIXEL_L: return GL_LUMINANCE;
|
|
case RS_KIND_PIXEL_A: return GL_ALPHA;
|
|
case RS_KIND_PIXEL_LA: return GL_LUMINANCE_ALPHA;
|
|
case RS_KIND_PIXEL_RGB: return GL_RGB;
|
|
case RS_KIND_PIXEL_RGBA: return GL_RGBA;
|
|
case RS_KIND_PIXEL_DEPTH: return GL_DEPTH_COMPONENT16;
|
|
default: break;
|
|
}
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
uint8_t *GetOffsetPtr(const android::renderscript::Allocation *alloc,
|
|
uint32_t xoff, uint32_t yoff, uint32_t zoff,
|
|
uint32_t lod, RsAllocationCubemapFace face) {
|
|
uint8_t *ptr = (uint8_t *)alloc->mHal.drvState.lod[lod].mallocPtr;
|
|
ptr += face * alloc->mHal.drvState.faceOffset;
|
|
ptr += zoff * alloc->mHal.drvState.lod[lod].dimY * alloc->mHal.drvState.lod[lod].stride;
|
|
ptr += yoff * alloc->mHal.drvState.lod[lod].stride;
|
|
ptr += xoff * alloc->mHal.state.elementSizeBytes;
|
|
return ptr;
|
|
}
|
|
|
|
|
|
static void Update2DTexture(const Context *rsc, const Allocation *alloc, const void *ptr,
|
|
uint32_t xoff, uint32_t yoff, uint32_t lod,
|
|
RsAllocationCubemapFace face, uint32_t w, uint32_t h) {
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
rsAssert(drv->textureID);
|
|
RSD_CALL_GL(glBindTexture, drv->glTarget, drv->textureID);
|
|
RSD_CALL_GL(glPixelStorei, GL_UNPACK_ALIGNMENT, 1);
|
|
GLenum t = GL_TEXTURE_2D;
|
|
if (alloc->mHal.state.hasFaces) {
|
|
t = gFaceOrder[face];
|
|
}
|
|
RSD_CALL_GL(glTexSubImage2D, t, lod, xoff, yoff, w, h, drv->glFormat, drv->glType, ptr);
|
|
#endif
|
|
}
|
|
|
|
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
static void Upload2DTexture(const Context *rsc, const Allocation *alloc, bool isFirstUpload) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
RSD_CALL_GL(glBindTexture, drv->glTarget, drv->textureID);
|
|
RSD_CALL_GL(glPixelStorei, GL_UNPACK_ALIGNMENT, 1);
|
|
|
|
uint32_t faceCount = 1;
|
|
if (alloc->mHal.state.hasFaces) {
|
|
faceCount = 6;
|
|
}
|
|
|
|
rsdGLCheckError(rsc, "Upload2DTexture 1 ");
|
|
for (uint32_t face = 0; face < faceCount; face ++) {
|
|
for (uint32_t lod = 0; lod < alloc->mHal.state.type->getLODCount(); lod++) {
|
|
const uint8_t *p = GetOffsetPtr(alloc, 0, 0, 0, lod, (RsAllocationCubemapFace)face);
|
|
|
|
GLenum t = GL_TEXTURE_2D;
|
|
if (alloc->mHal.state.hasFaces) {
|
|
t = gFaceOrder[face];
|
|
}
|
|
|
|
if (isFirstUpload) {
|
|
RSD_CALL_GL(glTexImage2D, t, lod, drv->glFormat,
|
|
alloc->mHal.state.type->getLODDimX(lod),
|
|
alloc->mHal.state.type->getLODDimY(lod),
|
|
0, drv->glFormat, drv->glType, p);
|
|
} else {
|
|
RSD_CALL_GL(glTexSubImage2D, t, lod, 0, 0,
|
|
alloc->mHal.state.type->getLODDimX(lod),
|
|
alloc->mHal.state.type->getLODDimY(lod),
|
|
drv->glFormat, drv->glType, p);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (alloc->mHal.state.mipmapControl == RS_ALLOCATION_MIPMAP_ON_SYNC_TO_TEXTURE) {
|
|
RSD_CALL_GL(glGenerateMipmap, drv->glTarget);
|
|
}
|
|
rsdGLCheckError(rsc, "Upload2DTexture");
|
|
}
|
|
#endif
|
|
|
|
static void UploadToTexture(const Context *rsc, const Allocation *alloc) {
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_INPUT) {
|
|
if (!drv->textureID) {
|
|
RSD_CALL_GL(glGenTextures, 1, &drv->textureID);
|
|
}
|
|
return;
|
|
}
|
|
|
|
if (!drv->glType || !drv->glFormat) {
|
|
return;
|
|
}
|
|
|
|
if (!alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
return;
|
|
}
|
|
|
|
bool isFirstUpload = false;
|
|
|
|
if (!drv->textureID) {
|
|
RSD_CALL_GL(glGenTextures, 1, &drv->textureID);
|
|
isFirstUpload = true;
|
|
}
|
|
|
|
Upload2DTexture(rsc, alloc, isFirstUpload);
|
|
|
|
if (!(alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_SCRIPT)) {
|
|
if (alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
free(alloc->mHal.drvState.lod[0].mallocPtr);
|
|
alloc->mHal.drvState.lod[0].mallocPtr = nullptr;
|
|
}
|
|
}
|
|
rsdGLCheckError(rsc, "UploadToTexture");
|
|
#endif
|
|
}
|
|
|
|
static void AllocateRenderTarget(const Context *rsc, const Allocation *alloc) {
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
if (!drv->glFormat) {
|
|
return;
|
|
}
|
|
|
|
if (!drv->renderTargetID) {
|
|
RSD_CALL_GL(glGenRenderbuffers, 1, &drv->renderTargetID);
|
|
|
|
if (!drv->renderTargetID) {
|
|
// This should generally not happen
|
|
ALOGE("allocateRenderTarget failed to gen mRenderTargetID");
|
|
rsc->dumpDebug();
|
|
return;
|
|
}
|
|
RSD_CALL_GL(glBindRenderbuffer, GL_RENDERBUFFER, drv->renderTargetID);
|
|
RSD_CALL_GL(glRenderbufferStorage, GL_RENDERBUFFER, drv->glFormat,
|
|
alloc->mHal.drvState.lod[0].dimX, alloc->mHal.drvState.lod[0].dimY);
|
|
}
|
|
rsdGLCheckError(rsc, "AllocateRenderTarget");
|
|
#endif
|
|
}
|
|
|
|
static void UploadToBufferObject(const Context *rsc, const Allocation *alloc) {
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
rsAssert(!alloc->mHal.state.type->getDimY());
|
|
rsAssert(!alloc->mHal.state.type->getDimZ());
|
|
|
|
//alloc->mHal.state.usageFlags |= RS_ALLOCATION_USAGE_GRAPHICS_VERTEX;
|
|
|
|
if (!drv->bufferID) {
|
|
RSD_CALL_GL(glGenBuffers, 1, &drv->bufferID);
|
|
}
|
|
if (!drv->bufferID) {
|
|
ALOGE("Upload to buffer object failed");
|
|
drv->uploadDeferred = true;
|
|
return;
|
|
}
|
|
RSD_CALL_GL(glBindBuffer, drv->glTarget, drv->bufferID);
|
|
RSD_CALL_GL(glBufferData, drv->glTarget,
|
|
alloc->mHal.state.type->getPackedSizeBytes(),
|
|
alloc->mHal.drvState.lod[0].mallocPtr, GL_DYNAMIC_DRAW);
|
|
RSD_CALL_GL(glBindBuffer, drv->glTarget, 0);
|
|
rsdGLCheckError(rsc, "UploadToBufferObject");
|
|
#endif
|
|
}
|
|
|
|
|
|
static size_t DeriveYUVLayout(int yuv, Allocation::Hal::DrvState *state) {
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
// For the flexible YCbCr format, layout is initialized during call to
|
|
// Allocation::ioReceive. Return early and avoid clobberring any
|
|
// pre-existing layout.
|
|
if (yuv == RS_YUV_420_888) {
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
// YUV only supports basic 2d
|
|
// so we can stash the plane pointers in the mipmap levels.
|
|
size_t uvSize = 0;
|
|
state->lod[1].dimX = state->lod[0].dimX / 2;
|
|
state->lod[1].dimY = state->lod[0].dimY / 2;
|
|
state->lod[2].dimX = state->lod[0].dimX / 2;
|
|
state->lod[2].dimY = state->lod[0].dimY / 2;
|
|
state->yuv.shift = 1;
|
|
state->yuv.step = 1;
|
|
state->lodCount = 3;
|
|
|
|
switch(yuv) {
|
|
case RS_YUV_YV12:
|
|
state->lod[2].stride = rsRound(state->lod[0].stride >> 1, 16);
|
|
state->lod[2].mallocPtr = ((uint8_t *)state->lod[0].mallocPtr) +
|
|
(state->lod[0].stride * state->lod[0].dimY);
|
|
uvSize += state->lod[2].stride * state->lod[2].dimY;
|
|
|
|
state->lod[1].stride = state->lod[2].stride;
|
|
state->lod[1].mallocPtr = ((uint8_t *)state->lod[2].mallocPtr) +
|
|
(state->lod[2].stride * state->lod[2].dimY);
|
|
uvSize += state->lod[1].stride * state->lod[2].dimY;
|
|
break;
|
|
case RS_YUV_NV21:
|
|
//state->lod[1].dimX = state->lod[0].dimX;
|
|
state->lod[1].stride = state->lod[0].stride;
|
|
state->lod[2].stride = state->lod[0].stride;
|
|
state->lod[2].mallocPtr = ((uint8_t *)state->lod[0].mallocPtr) +
|
|
(state->lod[0].stride * state->lod[0].dimY);
|
|
state->lod[1].mallocPtr = ((uint8_t *)state->lod[2].mallocPtr) + 1;
|
|
uvSize += state->lod[1].stride * state->lod[1].dimY;
|
|
state->yuv.step = 2;
|
|
break;
|
|
default:
|
|
rsAssert(0);
|
|
}
|
|
return uvSize;
|
|
}
|
|
|
|
static size_t AllocationBuildPointerTable(const Context *rsc, const Allocation *alloc,
|
|
const Type *type, uint8_t *ptr, size_t requiredAlignment) {
|
|
alloc->mHal.drvState.lod[0].dimX = type->getDimX();
|
|
alloc->mHal.drvState.lod[0].dimY = type->getDimY();
|
|
alloc->mHal.drvState.lod[0].dimZ = type->getDimZ();
|
|
alloc->mHal.drvState.lod[0].mallocPtr = 0;
|
|
// Stride needs to be aligned to a boundary defined by requiredAlignment!
|
|
size_t stride = alloc->mHal.drvState.lod[0].dimX * type->getElementSizeBytes();
|
|
alloc->mHal.drvState.lod[0].stride = rsRound(stride, requiredAlignment);
|
|
alloc->mHal.drvState.lodCount = type->getLODCount();
|
|
alloc->mHal.drvState.faceCount = type->getDimFaces();
|
|
|
|
size_t offsets[Allocation::MAX_LOD];
|
|
memset(offsets, 0, sizeof(offsets));
|
|
|
|
size_t o = alloc->mHal.drvState.lod[0].stride * rsMax(alloc->mHal.drvState.lod[0].dimY, 1u) *
|
|
rsMax(alloc->mHal.drvState.lod[0].dimZ, 1u);
|
|
if (alloc->mHal.state.yuv) {
|
|
o += DeriveYUVLayout(alloc->mHal.state.yuv, &alloc->mHal.drvState);
|
|
|
|
for (uint32_t ct = 1; ct < alloc->mHal.drvState.lodCount; ct++) {
|
|
offsets[ct] = (size_t)alloc->mHal.drvState.lod[ct].mallocPtr;
|
|
}
|
|
} else if(alloc->mHal.drvState.lodCount > 1) {
|
|
uint32_t tx = alloc->mHal.drvState.lod[0].dimX;
|
|
uint32_t ty = alloc->mHal.drvState.lod[0].dimY;
|
|
uint32_t tz = alloc->mHal.drvState.lod[0].dimZ;
|
|
for (uint32_t lod=1; lod < alloc->mHal.drvState.lodCount; lod++) {
|
|
alloc->mHal.drvState.lod[lod].dimX = tx;
|
|
alloc->mHal.drvState.lod[lod].dimY = ty;
|
|
alloc->mHal.drvState.lod[lod].dimZ = tz;
|
|
alloc->mHal.drvState.lod[lod].stride =
|
|
rsRound(tx * type->getElementSizeBytes(), requiredAlignment);
|
|
offsets[lod] = o;
|
|
o += alloc->mHal.drvState.lod[lod].stride * rsMax(ty, 1u) * rsMax(tz, 1u);
|
|
if (tx > 1) tx >>= 1;
|
|
if (ty > 1) ty >>= 1;
|
|
if (tz > 1) tz >>= 1;
|
|
}
|
|
}
|
|
|
|
alloc->mHal.drvState.faceOffset = o;
|
|
|
|
alloc->mHal.drvState.lod[0].mallocPtr = ptr;
|
|
for (uint32_t lod=1; lod < alloc->mHal.drvState.lodCount; lod++) {
|
|
alloc->mHal.drvState.lod[lod].mallocPtr = ptr + offsets[lod];
|
|
}
|
|
|
|
size_t allocSize = alloc->mHal.drvState.faceOffset;
|
|
if(alloc->mHal.drvState.faceCount) {
|
|
allocSize *= 6;
|
|
}
|
|
|
|
return allocSize;
|
|
}
|
|
|
|
static size_t AllocationBuildPointerTable(const Context *rsc, const Allocation *alloc,
|
|
const Type *type, uint8_t *ptr) {
|
|
return AllocationBuildPointerTable(rsc, alloc, type, ptr, Allocation::kMinimumRSAlignment);
|
|
}
|
|
|
|
static uint8_t* allocAlignedMemory(size_t allocSize, bool forceZero, size_t requiredAlignment) {
|
|
// We align all allocations to a boundary defined by requiredAlignment.
|
|
uint8_t* ptr = (uint8_t *)memalign(requiredAlignment, allocSize);
|
|
if (!ptr) {
|
|
return nullptr;
|
|
}
|
|
if (forceZero) {
|
|
memset(ptr, 0, allocSize);
|
|
}
|
|
return ptr;
|
|
}
|
|
|
|
bool rsdAllocationInitStrided(const Context *rsc, Allocation *alloc, bool forceZero, size_t requiredAlignment) {
|
|
DrvAllocation *drv = (DrvAllocation *)calloc(1, sizeof(DrvAllocation));
|
|
if (!drv) {
|
|
return false;
|
|
}
|
|
alloc->mHal.drv = drv;
|
|
|
|
// Check if requiredAlignment is power of 2, also requiredAlignment should be larger or equal than kMinimumRSAlignment.
|
|
if ((requiredAlignment & (requiredAlignment-1)) != 0 || requiredAlignment < Allocation::kMinimumRSAlignment) {
|
|
ALOGE("requiredAlignment must be power of 2");
|
|
return false;
|
|
}
|
|
// Calculate the object size.
|
|
size_t allocSize = AllocationBuildPointerTable(rsc, alloc, alloc->getType(), nullptr, requiredAlignment);
|
|
|
|
uint8_t * ptr = nullptr;
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_OUTPUT) {
|
|
|
|
} else if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_INPUT) {
|
|
// Allocation is allocated when the surface is created
|
|
// in getSurface
|
|
#ifdef RS_COMPATIBILITY_LIB
|
|
} else if (alloc->mHal.state.usageFlags == (RS_ALLOCATION_USAGE_INCREMENTAL_SUPPORT | RS_ALLOCATION_USAGE_SHARED)) {
|
|
if (alloc->mHal.state.userProvidedPtr == nullptr) {
|
|
ALOGE("User-backed buffer pointer cannot be null");
|
|
return false;
|
|
}
|
|
if (alloc->getType()->getDimLOD() || alloc->getType()->getDimFaces()) {
|
|
ALOGE("User-allocated buffers must not have multiple faces or LODs");
|
|
return false;
|
|
}
|
|
|
|
drv->useUserProvidedPtr = true;
|
|
ptr = (uint8_t*)alloc->mHal.state.userProvidedPtr;
|
|
#endif
|
|
} else if (alloc->mHal.state.userProvidedPtr != nullptr) {
|
|
// user-provided allocation
|
|
// limitations: no faces, no LOD, USAGE_SCRIPT or SCRIPT+TEXTURE only
|
|
if (!(alloc->mHal.state.usageFlags == (RS_ALLOCATION_USAGE_SCRIPT | RS_ALLOCATION_USAGE_SHARED) ||
|
|
alloc->mHal.state.usageFlags == (RS_ALLOCATION_USAGE_SCRIPT | RS_ALLOCATION_USAGE_SHARED | RS_ALLOCATION_USAGE_GRAPHICS_TEXTURE))) {
|
|
ALOGE("Can't use user-allocated buffers if usage is not USAGE_SCRIPT | USAGE_SHARED or USAGE_SCRIPT | USAGE_SHARED | USAGE_GRAPHICS_TEXTURE");
|
|
return false;
|
|
}
|
|
if (alloc->getType()->getDimLOD() || alloc->getType()->getDimFaces()) {
|
|
ALOGE("User-allocated buffers must not have multiple faces or LODs");
|
|
return false;
|
|
}
|
|
|
|
// rows must be aligned based on requiredAlignment.
|
|
// validate that here, otherwise fall back to not use the user-backed allocation
|
|
if (((alloc->getType()->getDimX() * alloc->getType()->getElement()->getSizeBytes()) % requiredAlignment) != 0) {
|
|
ALOGV("User-backed allocation failed stride requirement, falling back to separate allocation");
|
|
drv->useUserProvidedPtr = false;
|
|
|
|
ptr = allocAlignedMemory(allocSize, forceZero, requiredAlignment);
|
|
if (!ptr) {
|
|
alloc->mHal.drv = nullptr;
|
|
free(drv);
|
|
return false;
|
|
}
|
|
|
|
} else {
|
|
drv->useUserProvidedPtr = true;
|
|
ptr = (uint8_t*)alloc->mHal.state.userProvidedPtr;
|
|
}
|
|
} else {
|
|
ptr = allocAlignedMemory(allocSize, forceZero, requiredAlignment);
|
|
if (!ptr) {
|
|
alloc->mHal.drv = nullptr;
|
|
free(drv);
|
|
return false;
|
|
}
|
|
}
|
|
// Build the pointer tables
|
|
size_t verifySize = AllocationBuildPointerTable(rsc, alloc, alloc->getType(), ptr, requiredAlignment);
|
|
if(allocSize != verifySize) {
|
|
rsAssert(!"Size mismatch");
|
|
}
|
|
|
|
drv->glTarget = GL_NONE;
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_GRAPHICS_TEXTURE) {
|
|
if (alloc->mHal.state.hasFaces) {
|
|
drv->glTarget = GL_TEXTURE_CUBE_MAP;
|
|
} else {
|
|
drv->glTarget = GL_TEXTURE_2D;
|
|
}
|
|
} else {
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_GRAPHICS_VERTEX) {
|
|
drv->glTarget = GL_ARRAY_BUFFER;
|
|
}
|
|
}
|
|
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
drv->glType = rsdTypeToGLType(alloc->mHal.state.type->getElement()->getComponent().getType());
|
|
drv->glFormat = rsdKindToGLFormat(alloc->mHal.state.type->getElement()->getComponent().getKind());
|
|
#else
|
|
drv->glType = 0;
|
|
drv->glFormat = 0;
|
|
#endif
|
|
|
|
if (alloc->mHal.state.usageFlags & ~RS_ALLOCATION_USAGE_SCRIPT) {
|
|
drv->uploadDeferred = true;
|
|
}
|
|
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
drv->readBackFBO = nullptr;
|
|
#endif
|
|
|
|
// fill out the initial state of the buffer if we couldn't use the user-provided ptr and USAGE_SHARED was accepted
|
|
if ((alloc->mHal.state.userProvidedPtr != 0) && (drv->useUserProvidedPtr == false)) {
|
|
rsdAllocationData2D(rsc, alloc, 0, 0, 0, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X, alloc->getType()->getDimX(), alloc->getType()->getDimY(), alloc->mHal.state.userProvidedPtr, allocSize, 0);
|
|
}
|
|
|
|
|
|
#ifdef RS_FIND_OFFSETS
|
|
ALOGE("pointer for allocation: %p", alloc);
|
|
ALOGE("pointer for allocation.drv: %p", &alloc->mHal.drv);
|
|
#endif
|
|
|
|
|
|
return true;
|
|
}
|
|
|
|
bool rsdAllocationInit(const Context *rsc, Allocation *alloc, bool forceZero) {
|
|
return rsdAllocationInitStrided(rsc, alloc, forceZero, Allocation::kMinimumRSAlignment);
|
|
}
|
|
|
|
void rsdAllocationAdapterOffset(const Context *rsc, const Allocation *alloc) {
|
|
//ALOGE("rsdAllocationAdapterOffset");
|
|
|
|
// Get a base pointer to the new LOD
|
|
const Allocation *base = alloc->mHal.state.baseAlloc;
|
|
const Type *type = alloc->mHal.state.type;
|
|
if (base == nullptr) {
|
|
return;
|
|
}
|
|
|
|
//ALOGE("rsdAllocationAdapterOffset %p %p", ptrA, ptrB);
|
|
//ALOGE("rsdAllocationAdapterOffset lodCount %i", alloc->mHal.drvState.lodCount);
|
|
|
|
const int lodBias = alloc->mHal.state.originLOD;
|
|
uint32_t lodCount = rsMax(alloc->mHal.drvState.lodCount, (uint32_t)1);
|
|
for (uint32_t lod=0; lod < lodCount; lod++) {
|
|
alloc->mHal.drvState.lod[lod] = base->mHal.drvState.lod[lod + lodBias];
|
|
alloc->mHal.drvState.lod[lod].mallocPtr = GetOffsetPtr(alloc,
|
|
alloc->mHal.state.originX, alloc->mHal.state.originY, alloc->mHal.state.originZ,
|
|
lodBias, (RsAllocationCubemapFace)alloc->mHal.state.originFace);
|
|
}
|
|
}
|
|
|
|
bool rsdAllocationAdapterInit(const Context *rsc, Allocation *alloc) {
|
|
DrvAllocation *drv = (DrvAllocation *)calloc(1, sizeof(DrvAllocation));
|
|
if (!drv) {
|
|
return false;
|
|
}
|
|
alloc->mHal.drv = drv;
|
|
|
|
// We need to build an allocation that looks like a subset of the parent allocation
|
|
rsdAllocationAdapterOffset(rsc, alloc);
|
|
|
|
return true;
|
|
}
|
|
|
|
void rsdAllocationDestroy(const Context *rsc, Allocation *alloc) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
if (alloc->mHal.state.baseAlloc == nullptr) {
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
if (drv->bufferID) {
|
|
// Causes a SW crash....
|
|
//ALOGV(" mBufferID %i", mBufferID);
|
|
//glDeleteBuffers(1, &mBufferID);
|
|
//mBufferID = 0;
|
|
}
|
|
if (drv->textureID) {
|
|
RSD_CALL_GL(glDeleteTextures, 1, &drv->textureID);
|
|
drv->textureID = 0;
|
|
}
|
|
if (drv->renderTargetID) {
|
|
RSD_CALL_GL(glDeleteRenderbuffers, 1, &drv->renderTargetID);
|
|
drv->renderTargetID = 0;
|
|
}
|
|
#endif
|
|
|
|
if (alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
// don't free user-allocated ptrs or IO_OUTPUT buffers
|
|
if (!(drv->useUserProvidedPtr) &&
|
|
!(alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_INPUT) &&
|
|
!(alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_OUTPUT)) {
|
|
free(alloc->mHal.drvState.lod[0].mallocPtr);
|
|
}
|
|
alloc->mHal.drvState.lod[0].mallocPtr = nullptr;
|
|
}
|
|
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
#ifndef RS_VENDOR_LIB
|
|
if (drv->readBackFBO != nullptr) {
|
|
delete drv->readBackFBO;
|
|
drv->readBackFBO = nullptr;
|
|
}
|
|
#endif
|
|
if ((alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_OUTPUT) &&
|
|
(alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_SCRIPT)) {
|
|
ANativeWindow *nw = drv->wndSurface;
|
|
if (nw) {
|
|
//If we have an attached surface, need to release it.
|
|
AHardwareBuffer* ahwb = ANativeWindowBuffer_getHardwareBuffer(drv->wndBuffer);
|
|
int fenceID = -1;
|
|
AHardwareBuffer_unlock(ahwb, &fenceID);
|
|
ANativeWindow_cancelBuffer(nw, drv->wndBuffer, fenceID);
|
|
ANativeWindow_release(nw);
|
|
drv->wndSurface = nullptr;
|
|
drv->wndBuffer = nullptr;
|
|
}
|
|
}
|
|
#endif
|
|
}
|
|
|
|
free(drv);
|
|
alloc->mHal.drv = nullptr;
|
|
}
|
|
|
|
void rsdAllocationResize(const Context *rsc, const Allocation *alloc,
|
|
const Type *newType, bool zeroNew) {
|
|
const uint32_t oldDimX = alloc->mHal.drvState.lod[0].dimX;
|
|
const uint32_t dimX = newType->getDimX();
|
|
|
|
// can't resize Allocations with user-allocated buffers
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_SHARED) {
|
|
ALOGE("Resize cannot be called on a USAGE_SHARED allocation");
|
|
return;
|
|
}
|
|
void * oldPtr = alloc->mHal.drvState.lod[0].mallocPtr;
|
|
// Calculate the object size
|
|
size_t s = AllocationBuildPointerTable(rsc, alloc, newType, nullptr);
|
|
uint8_t *ptr = (uint8_t *)realloc(oldPtr, s);
|
|
// Build the relative pointer tables.
|
|
size_t verifySize = AllocationBuildPointerTable(rsc, alloc, newType, ptr);
|
|
if(s != verifySize) {
|
|
rsAssert(!"Size mismatch");
|
|
}
|
|
|
|
|
|
if (dimX > oldDimX) {
|
|
size_t stride = alloc->mHal.state.elementSizeBytes;
|
|
memset(((uint8_t *)alloc->mHal.drvState.lod[0].mallocPtr) + stride * oldDimX,
|
|
0, stride * (dimX - oldDimX));
|
|
}
|
|
}
|
|
|
|
static void rsdAllocationSyncFromFBO(const Context *rsc, const Allocation *alloc) {
|
|
#if !defined(RS_VENDOR_LIB) && !defined(RS_COMPATIBILITY_LIB)
|
|
if (!alloc->getIsScript()) {
|
|
return; // nothing to sync
|
|
}
|
|
|
|
RsdHal *dc = (RsdHal *)rsc->mHal.drv;
|
|
RsdFrameBufferObj *lastFbo = dc->gl.currentFrameBuffer;
|
|
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
if (!drv->textureID && !drv->renderTargetID) {
|
|
return; // nothing was rendered here yet, so nothing to sync
|
|
}
|
|
if (drv->readBackFBO == nullptr) {
|
|
drv->readBackFBO = new RsdFrameBufferObj();
|
|
drv->readBackFBO->setColorTarget(drv, 0);
|
|
drv->readBackFBO->setDimensions(alloc->getType()->getDimX(),
|
|
alloc->getType()->getDimY());
|
|
}
|
|
|
|
// Bind the framebuffer object so we can read back from it
|
|
drv->readBackFBO->setActive(rsc);
|
|
|
|
// Do the readback
|
|
RSD_CALL_GL(glReadPixels, 0, 0, alloc->mHal.drvState.lod[0].dimX,
|
|
alloc->mHal.drvState.lod[0].dimY,
|
|
drv->glFormat, drv->glType, alloc->mHal.drvState.lod[0].mallocPtr);
|
|
|
|
// Revert framebuffer to its original
|
|
lastFbo->setActive(rsc);
|
|
#endif
|
|
}
|
|
|
|
|
|
void rsdAllocationSyncAll(const Context *rsc, const Allocation *alloc,
|
|
RsAllocationUsageType src) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
if (src == RS_ALLOCATION_USAGE_GRAPHICS_RENDER_TARGET) {
|
|
if(!alloc->getIsRenderTarget()) {
|
|
rsc->setError(RS_ERROR_FATAL_DRIVER,
|
|
"Attempting to sync allocation from render target, "
|
|
"for non-render target allocation");
|
|
} else if (alloc->getType()->getElement()->getKind() != RS_KIND_PIXEL_RGBA) {
|
|
rsc->setError(RS_ERROR_FATAL_DRIVER, "Cannot only sync from RGBA"
|
|
"render target");
|
|
} else {
|
|
rsdAllocationSyncFromFBO(rsc, alloc);
|
|
}
|
|
return;
|
|
}
|
|
|
|
rsAssert(src == RS_ALLOCATION_USAGE_SCRIPT || src == RS_ALLOCATION_USAGE_SHARED);
|
|
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_GRAPHICS_TEXTURE) {
|
|
UploadToTexture(rsc, alloc);
|
|
} else {
|
|
if ((alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_GRAPHICS_RENDER_TARGET) &&
|
|
!(alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_IO_OUTPUT)) {
|
|
AllocateRenderTarget(rsc, alloc);
|
|
}
|
|
}
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_GRAPHICS_VERTEX) {
|
|
UploadToBufferObject(rsc, alloc);
|
|
}
|
|
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_SHARED) {
|
|
|
|
if (src == RS_ALLOCATION_USAGE_SHARED) {
|
|
// just a memory fence for the CPU driver
|
|
// vendor drivers probably want to flush any dirty cachelines for
|
|
// this particular Allocation
|
|
__sync_synchronize();
|
|
}
|
|
}
|
|
|
|
drv->uploadDeferred = false;
|
|
}
|
|
|
|
void rsdAllocationMarkDirty(const Context *rsc, const Allocation *alloc) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
drv->uploadDeferred = true;
|
|
}
|
|
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
static bool IoGetBuffer(const Context *rsc, Allocation *alloc, ANativeWindow *nw) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
// Must lock the whole surface
|
|
int fenceID = -1;
|
|
int r = ANativeWindow_dequeueBuffer(nw, &drv->wndBuffer, &fenceID);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error dequeueing IO output buffer.");
|
|
close(fenceID);
|
|
return false;
|
|
}
|
|
|
|
void *dst = nullptr;
|
|
AHardwareBuffer* ahwb = ANativeWindowBuffer_getHardwareBuffer(drv->wndBuffer);
|
|
r = AHardwareBuffer_lock(ahwb, AHARDWAREBUFFER_USAGE_CPU_WRITE_OFTEN,
|
|
fenceID, NULL, &dst);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error Locking IO output buffer.");
|
|
return false;
|
|
}
|
|
alloc->mHal.drvState.lod[0].mallocPtr = dst;
|
|
alloc->mHal.drvState.lod[0].stride = drv->wndBuffer->stride * alloc->mHal.state.elementSizeBytes;
|
|
rsAssert((alloc->mHal.drvState.lod[0].stride & 0xf) == 0);
|
|
|
|
return true;
|
|
}
|
|
#endif
|
|
|
|
void rsdAllocationSetSurface(const Context *rsc, Allocation *alloc, ANativeWindow *nw) {
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
// Cleanup old surface if there is one.
|
|
if (drv->wndSurface) {
|
|
ANativeWindow *old = drv->wndSurface;
|
|
AHardwareBuffer* ahwb = ANativeWindowBuffer_getHardwareBuffer(drv->wndBuffer);
|
|
int fenceID = -1;
|
|
int32_t r = AHardwareBuffer_unlock(ahwb, &fenceID);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error unlocking output buffer.");
|
|
close(fenceID);
|
|
return;
|
|
}
|
|
r = ANativeWindow_cancelBuffer(old, drv->wndBuffer, fenceID);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error canceling output buffer.");
|
|
return;
|
|
}
|
|
ANativeWindow_release(old);
|
|
drv->wndSurface = nullptr;
|
|
drv->wndBuffer = nullptr;
|
|
}
|
|
|
|
if (nw) {
|
|
int32_t r = ANativeWindow_setBuffersGeometry(nw, alloc->mHal.drvState.lod[0].dimX,
|
|
alloc->mHal.drvState.lod[0].dimY,
|
|
WINDOW_FORMAT_RGBA_8888);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error setting IO output buffer geometry.");
|
|
return;
|
|
}
|
|
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_SCRIPT) {
|
|
r = ANativeWindow_setUsage(nw,
|
|
AHARDWAREBUFFER_USAGE_CPU_READ_RARELY | AHARDWAREBUFFER_USAGE_CPU_WRITE_OFTEN);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error setting IO output buffer usage.");
|
|
return;
|
|
}
|
|
}
|
|
|
|
IoGetBuffer(rsc, alloc, nw);
|
|
drv->wndSurface = nw;
|
|
}
|
|
|
|
return;
|
|
#endif
|
|
}
|
|
|
|
void rsdAllocationIoSend(const Context *rsc, Allocation *alloc) {
|
|
#ifndef RS_COMPATIBILITY_LIB
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
ANativeWindow *nw = drv->wndSurface;
|
|
#ifndef RS_VENDOR_LIB
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_GRAPHICS_RENDER_TARGET) {
|
|
RsdHal *dc = (RsdHal *)rsc->mHal.drv;
|
|
RSD_CALL_GL(eglSwapBuffers, dc->gl.egl.display, dc->gl.egl.surface);
|
|
return;
|
|
}
|
|
#endif
|
|
if (nw) {
|
|
if (alloc->mHal.state.usageFlags & RS_ALLOCATION_USAGE_SCRIPT) {
|
|
AHardwareBuffer* ahwb = ANativeWindowBuffer_getHardwareBuffer(drv->wndBuffer);
|
|
int fenceID = -1;
|
|
int32_t r = AHardwareBuffer_unlock(ahwb, &fenceID);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error unlock output buffer.");
|
|
close(fenceID);
|
|
return;
|
|
}
|
|
r = ANativeWindow_queueBuffer(nw, drv->wndBuffer, fenceID);
|
|
if (r) {
|
|
rsc->setError(RS_ERROR_DRIVER, "Error sending IO output buffer.");
|
|
return;
|
|
}
|
|
drv->wndBuffer = nullptr;
|
|
IoGetBuffer(rsc, alloc, nw);
|
|
}
|
|
} else {
|
|
rsc->setError(RS_ERROR_DRIVER, "Sent IO buffer with no attached surface.");
|
|
return;
|
|
}
|
|
#endif
|
|
}
|
|
|
|
void rsdAllocationIoReceive(const Context *rsc, Allocation *alloc) {
|
|
if (alloc->mHal.state.yuv) {
|
|
DeriveYUVLayout(alloc->mHal.state.yuv, &alloc->mHal.drvState);
|
|
}
|
|
}
|
|
|
|
|
|
void rsdAllocationData1D(const Context *rsc, const Allocation *alloc,
|
|
uint32_t xoff, uint32_t lod, size_t count,
|
|
const void *data, size_t sizeBytes) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
const size_t eSize = alloc->mHal.state.type->getElementSizeBytes();
|
|
uint8_t * ptr = GetOffsetPtr(alloc, xoff, 0, 0, 0, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
size_t size = count * eSize;
|
|
if (ptr != data) {
|
|
// Skip the copy if we are the same allocation. This can arise from
|
|
// our Bitmap optimization, where we share the same storage.
|
|
if (alloc->mHal.state.hasReferences) {
|
|
alloc->incRefs(data, count);
|
|
alloc->decRefs(ptr, count);
|
|
}
|
|
memcpy(ptr, data, size);
|
|
}
|
|
drv->uploadDeferred = true;
|
|
}
|
|
|
|
void rsdAllocationData2D(const Context *rsc, const Allocation *alloc,
|
|
uint32_t xoff, uint32_t yoff, uint32_t lod, RsAllocationCubemapFace face,
|
|
uint32_t w, uint32_t h, const void *data, size_t sizeBytes, size_t stride) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
size_t eSize = alloc->mHal.state.elementSizeBytes;
|
|
size_t lineSize = eSize * w;
|
|
if (!stride) {
|
|
stride = lineSize;
|
|
}
|
|
|
|
if (alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
const uint8_t *src = static_cast<const uint8_t *>(data);
|
|
uint8_t *dst = GetOffsetPtr(alloc, xoff, yoff, 0, lod, face);
|
|
if (dst == src) {
|
|
// Skip the copy if we are the same allocation. This can arise from
|
|
// our Bitmap optimization, where we share the same storage.
|
|
drv->uploadDeferred = true;
|
|
return;
|
|
}
|
|
|
|
for (uint32_t line=yoff; line < (yoff+h); line++) {
|
|
if (alloc->mHal.state.hasReferences) {
|
|
alloc->incRefs(src, w);
|
|
alloc->decRefs(dst, w);
|
|
}
|
|
memcpy(dst, src, lineSize);
|
|
src += stride;
|
|
dst += alloc->mHal.drvState.lod[lod].stride;
|
|
}
|
|
if (alloc->mHal.state.yuv) {
|
|
size_t clineSize = lineSize;
|
|
int lod = 1;
|
|
int maxLod = 2;
|
|
if (alloc->mHal.state.yuv == RS_YUV_YV12) {
|
|
maxLod = 3;
|
|
clineSize >>= 1;
|
|
} else if (alloc->mHal.state.yuv == RS_YUV_NV21) {
|
|
lod = 2;
|
|
maxLod = 3;
|
|
}
|
|
|
|
while (lod < maxLod) {
|
|
uint8_t *dst = GetOffsetPtr(alloc, xoff, yoff, 0, lod, face);
|
|
|
|
for (uint32_t line=(yoff >> 1); line < ((yoff+h)>>1); line++) {
|
|
memcpy(dst, src, clineSize);
|
|
// When copying from an array to an Allocation, the src pointer
|
|
// to the array should just move by the number of bytes copied.
|
|
src += clineSize;
|
|
dst += alloc->mHal.drvState.lod[lod].stride;
|
|
}
|
|
lod++;
|
|
}
|
|
|
|
}
|
|
drv->uploadDeferred = true;
|
|
} else {
|
|
Update2DTexture(rsc, alloc, data, xoff, yoff, lod, face, w, h);
|
|
}
|
|
}
|
|
|
|
void rsdAllocationData3D(const Context *rsc, const Allocation *alloc,
|
|
uint32_t xoff, uint32_t yoff, uint32_t zoff,
|
|
uint32_t lod,
|
|
uint32_t w, uint32_t h, uint32_t d, const void *data,
|
|
size_t sizeBytes, size_t stride) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
uint32_t eSize = alloc->mHal.state.elementSizeBytes;
|
|
uint32_t lineSize = eSize * w;
|
|
if (!stride) {
|
|
stride = lineSize;
|
|
}
|
|
|
|
if (alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
const uint8_t *src = static_cast<const uint8_t *>(data);
|
|
for (uint32_t z = zoff; z < (d + zoff); z++) {
|
|
uint8_t *dst = GetOffsetPtr(alloc, xoff, yoff, z, lod,
|
|
RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
if (dst == src) {
|
|
// Skip the copy if we are the same allocation. This can arise from
|
|
// our Bitmap optimization, where we share the same storage.
|
|
drv->uploadDeferred = true;
|
|
return;
|
|
}
|
|
|
|
for (uint32_t line=yoff; line < (yoff+h); line++) {
|
|
if (alloc->mHal.state.hasReferences) {
|
|
alloc->incRefs(src, w);
|
|
alloc->decRefs(dst, w);
|
|
}
|
|
memcpy(dst, src, lineSize);
|
|
src += stride;
|
|
dst += alloc->mHal.drvState.lod[lod].stride;
|
|
}
|
|
}
|
|
drv->uploadDeferred = true;
|
|
}
|
|
}
|
|
|
|
void rsdAllocationRead1D(const Context *rsc, const Allocation *alloc,
|
|
uint32_t xoff, uint32_t lod, size_t count,
|
|
void *data, size_t sizeBytes) {
|
|
const size_t eSize = alloc->mHal.state.type->getElementSizeBytes();
|
|
const uint8_t * ptr = GetOffsetPtr(alloc, xoff, 0, 0, 0, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
if (data != ptr) {
|
|
// Skip the copy if we are the same allocation. This can arise from
|
|
// our Bitmap optimization, where we share the same storage.
|
|
memcpy(data, ptr, count * eSize);
|
|
}
|
|
}
|
|
|
|
void rsdAllocationRead2D(const Context *rsc, const Allocation *alloc,
|
|
uint32_t xoff, uint32_t yoff, uint32_t lod, RsAllocationCubemapFace face,
|
|
uint32_t w, uint32_t h, void *data, size_t sizeBytes, size_t stride) {
|
|
size_t eSize = alloc->mHal.state.elementSizeBytes;
|
|
size_t lineSize = eSize * w;
|
|
if (!stride) {
|
|
stride = lineSize;
|
|
}
|
|
|
|
if (alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
uint8_t *dst = static_cast<uint8_t *>(data);
|
|
const uint8_t *src = GetOffsetPtr(alloc, xoff, yoff, 0, lod, face);
|
|
if (dst == src) {
|
|
// Skip the copy if we are the same allocation. This can arise from
|
|
// our Bitmap optimization, where we share the same storage.
|
|
return;
|
|
}
|
|
|
|
for (uint32_t line=yoff; line < (yoff+h); line++) {
|
|
memcpy(dst, src, lineSize);
|
|
dst += stride;
|
|
src += alloc->mHal.drvState.lod[lod].stride;
|
|
}
|
|
} else {
|
|
ALOGE("Add code to readback from non-script memory");
|
|
}
|
|
}
|
|
|
|
|
|
void rsdAllocationRead3D(const Context *rsc, const Allocation *alloc,
|
|
uint32_t xoff, uint32_t yoff, uint32_t zoff,
|
|
uint32_t lod,
|
|
uint32_t w, uint32_t h, uint32_t d, void *data, size_t sizeBytes, size_t stride) {
|
|
uint32_t eSize = alloc->mHal.state.elementSizeBytes;
|
|
uint32_t lineSize = eSize * w;
|
|
if (!stride) {
|
|
stride = lineSize;
|
|
}
|
|
|
|
if (alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
uint8_t *dst = static_cast<uint8_t *>(data);
|
|
for (uint32_t z = zoff; z < (d + zoff); z++) {
|
|
const uint8_t *src = GetOffsetPtr(alloc, xoff, yoff, z, lod,
|
|
RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
if (dst == src) {
|
|
// Skip the copy if we are the same allocation. This can arise from
|
|
// our Bitmap optimization, where we share the same storage.
|
|
return;
|
|
}
|
|
|
|
for (uint32_t line=yoff; line < (yoff+h); line++) {
|
|
memcpy(dst, src, lineSize);
|
|
dst += stride;
|
|
src += alloc->mHal.drvState.lod[lod].stride;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
void * rsdAllocationLock1D(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *alloc) {
|
|
return alloc->mHal.drvState.lod[0].mallocPtr;
|
|
}
|
|
|
|
void rsdAllocationUnlock1D(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *alloc) {
|
|
|
|
}
|
|
|
|
void rsdAllocationData1D_alloc(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *dstAlloc,
|
|
uint32_t dstXoff, uint32_t dstLod, size_t count,
|
|
const android::renderscript::Allocation *srcAlloc,
|
|
uint32_t srcXoff, uint32_t srcLod) {
|
|
}
|
|
|
|
|
|
void rsdAllocationData2D_alloc_script(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *dstAlloc,
|
|
uint32_t dstXoff, uint32_t dstYoff, uint32_t dstLod,
|
|
RsAllocationCubemapFace dstFace, uint32_t w, uint32_t h,
|
|
const android::renderscript::Allocation *srcAlloc,
|
|
uint32_t srcXoff, uint32_t srcYoff, uint32_t srcLod,
|
|
RsAllocationCubemapFace srcFace) {
|
|
size_t elementSize = dstAlloc->getType()->getElementSizeBytes();
|
|
for (uint32_t i = 0; i < h; i ++) {
|
|
uint8_t *dstPtr = GetOffsetPtr(dstAlloc, dstXoff, dstYoff + i, 0, dstLod, dstFace);
|
|
uint8_t *srcPtr = GetOffsetPtr(srcAlloc, srcXoff, srcYoff + i, 0, srcLod, srcFace);
|
|
memcpy(dstPtr, srcPtr, w * elementSize);
|
|
|
|
//ALOGE("COPIED dstXoff(%u), dstYoff(%u), dstLod(%u), dstFace(%u), w(%u), h(%u), srcXoff(%u), srcYoff(%u), srcLod(%u), srcFace(%u)",
|
|
// dstXoff, dstYoff, dstLod, dstFace, w, h, srcXoff, srcYoff, srcLod, srcFace);
|
|
}
|
|
}
|
|
|
|
void rsdAllocationData3D_alloc_script(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *dstAlloc,
|
|
uint32_t dstXoff, uint32_t dstYoff, uint32_t dstZoff, uint32_t dstLod,
|
|
uint32_t w, uint32_t h, uint32_t d,
|
|
const android::renderscript::Allocation *srcAlloc,
|
|
uint32_t srcXoff, uint32_t srcYoff, uint32_t srcZoff, uint32_t srcLod) {
|
|
uint32_t elementSize = dstAlloc->getType()->getElementSizeBytes();
|
|
for (uint32_t j = 0; j < d; j++) {
|
|
for (uint32_t i = 0; i < h; i ++) {
|
|
uint8_t *dstPtr = GetOffsetPtr(dstAlloc, dstXoff, dstYoff + i, dstZoff + j,
|
|
dstLod, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
uint8_t *srcPtr = GetOffsetPtr(srcAlloc, srcXoff, srcYoff + i, srcZoff + j,
|
|
srcLod, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
memcpy(dstPtr, srcPtr, w * elementSize);
|
|
|
|
//ALOGE("COPIED dstXoff(%u), dstYoff(%u), dstLod(%u), dstFace(%u), w(%u), h(%u), srcXoff(%u), srcYoff(%u), srcLod(%u), srcFace(%u)",
|
|
// dstXoff, dstYoff, dstLod, dstFace, w, h, srcXoff, srcYoff, srcLod, srcFace);
|
|
}
|
|
}
|
|
}
|
|
|
|
void rsdAllocationData2D_alloc(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *dstAlloc,
|
|
uint32_t dstXoff, uint32_t dstYoff, uint32_t dstLod,
|
|
RsAllocationCubemapFace dstFace, uint32_t w, uint32_t h,
|
|
const android::renderscript::Allocation *srcAlloc,
|
|
uint32_t srcXoff, uint32_t srcYoff, uint32_t srcLod,
|
|
RsAllocationCubemapFace srcFace) {
|
|
if (!dstAlloc->getIsScript() && !srcAlloc->getIsScript()) {
|
|
rsc->setError(RS_ERROR_FATAL_DRIVER, "Non-script allocation copies not "
|
|
"yet implemented.");
|
|
return;
|
|
}
|
|
rsdAllocationData2D_alloc_script(rsc, dstAlloc, dstXoff, dstYoff,
|
|
dstLod, dstFace, w, h, srcAlloc,
|
|
srcXoff, srcYoff, srcLod, srcFace);
|
|
}
|
|
|
|
void rsdAllocationData3D_alloc(const android::renderscript::Context *rsc,
|
|
const android::renderscript::Allocation *dstAlloc,
|
|
uint32_t dstXoff, uint32_t dstYoff, uint32_t dstZoff,
|
|
uint32_t dstLod,
|
|
uint32_t w, uint32_t h, uint32_t d,
|
|
const android::renderscript::Allocation *srcAlloc,
|
|
uint32_t srcXoff, uint32_t srcYoff, uint32_t srcZoff,
|
|
uint32_t srcLod) {
|
|
if (!dstAlloc->getIsScript() && !srcAlloc->getIsScript()) {
|
|
rsc->setError(RS_ERROR_FATAL_DRIVER, "Non-script allocation copies not "
|
|
"yet implemented.");
|
|
return;
|
|
}
|
|
rsdAllocationData3D_alloc_script(rsc, dstAlloc, dstXoff, dstYoff, dstZoff,
|
|
dstLod, w, h, d, srcAlloc,
|
|
srcXoff, srcYoff, srcZoff, srcLod);
|
|
}
|
|
|
|
void rsdAllocationElementData(const Context *rsc, const Allocation *alloc,
|
|
uint32_t x, uint32_t y, uint32_t z,
|
|
const void *data, uint32_t cIdx, size_t sizeBytes) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
uint8_t * ptr = GetOffsetPtr(alloc, x, y, z, 0, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
|
|
const Element * e = alloc->mHal.state.type->getElement()->getField(cIdx);
|
|
ptr += alloc->mHal.state.type->getElement()->getFieldOffsetBytes(cIdx);
|
|
|
|
if (alloc->mHal.state.hasReferences) {
|
|
e->incRefs(data);
|
|
e->decRefs(ptr);
|
|
}
|
|
|
|
memcpy(ptr, data, sizeBytes);
|
|
drv->uploadDeferred = true;
|
|
}
|
|
|
|
void rsdAllocationElementRead(const Context *rsc, const Allocation *alloc,
|
|
uint32_t x, uint32_t y, uint32_t z,
|
|
void *data, uint32_t cIdx, size_t sizeBytes) {
|
|
DrvAllocation *drv = (DrvAllocation *)alloc->mHal.drv;
|
|
|
|
uint8_t * ptr = GetOffsetPtr(alloc, x, y, z, 0, RS_ALLOCATION_CUBEMAP_FACE_POSITIVE_X);
|
|
|
|
const Element * e = alloc->mHal.state.type->getElement()->getField(cIdx);
|
|
ptr += alloc->mHal.state.type->getElement()->getFieldOffsetBytes(cIdx);
|
|
|
|
memcpy(data, ptr, sizeBytes);
|
|
}
|
|
|
|
static void mip565(const Allocation *alloc, int lod, RsAllocationCubemapFace face) {
|
|
uint32_t w = alloc->mHal.drvState.lod[lod + 1].dimX;
|
|
uint32_t h = alloc->mHal.drvState.lod[lod + 1].dimY;
|
|
|
|
for (uint32_t y=0; y < h; y++) {
|
|
uint16_t *oPtr = (uint16_t *)GetOffsetPtr(alloc, 0, y, 0, lod + 1, face);
|
|
const uint16_t *i1 = (uint16_t *)GetOffsetPtr(alloc, 0, 0, y*2, lod, face);
|
|
const uint16_t *i2 = (uint16_t *)GetOffsetPtr(alloc, 0, 0, y*2+1, lod, face);
|
|
|
|
for (uint32_t x=0; x < w; x++) {
|
|
*oPtr = rsBoxFilter565(i1[0], i1[1], i2[0], i2[1]);
|
|
oPtr ++;
|
|
i1 += 2;
|
|
i2 += 2;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void mip8888(const Allocation *alloc, int lod, RsAllocationCubemapFace face) {
|
|
uint32_t w = alloc->mHal.drvState.lod[lod + 1].dimX;
|
|
uint32_t h = alloc->mHal.drvState.lod[lod + 1].dimY;
|
|
|
|
for (uint32_t y=0; y < h; y++) {
|
|
uint32_t *oPtr = (uint32_t *)GetOffsetPtr(alloc, 0, y, 0, lod + 1, face);
|
|
const uint32_t *i1 = (uint32_t *)GetOffsetPtr(alloc, 0, y*2, 0, lod, face);
|
|
const uint32_t *i2 = (uint32_t *)GetOffsetPtr(alloc, 0, y*2+1, 0, lod, face);
|
|
|
|
for (uint32_t x=0; x < w; x++) {
|
|
*oPtr = rsBoxFilter8888(i1[0], i1[1], i2[0], i2[1]);
|
|
oPtr ++;
|
|
i1 += 2;
|
|
i2 += 2;
|
|
}
|
|
}
|
|
}
|
|
|
|
static void mip8(const Allocation *alloc, int lod, RsAllocationCubemapFace face) {
|
|
uint32_t w = alloc->mHal.drvState.lod[lod + 1].dimX;
|
|
uint32_t h = alloc->mHal.drvState.lod[lod + 1].dimY;
|
|
|
|
for (uint32_t y=0; y < h; y++) {
|
|
uint8_t *oPtr = GetOffsetPtr(alloc, 0, y, 0, lod + 1, face);
|
|
const uint8_t *i1 = GetOffsetPtr(alloc, 0, y*2, 0, lod, face);
|
|
const uint8_t *i2 = GetOffsetPtr(alloc, 0, y*2+1, 0, lod, face);
|
|
|
|
for (uint32_t x=0; x < w; x++) {
|
|
*oPtr = (uint8_t)(((uint32_t)i1[0] + i1[1] + i2[0] + i2[1]) * 0.25f);
|
|
oPtr ++;
|
|
i1 += 2;
|
|
i2 += 2;
|
|
}
|
|
}
|
|
}
|
|
|
|
void rsdAllocationGenerateMipmaps(const Context *rsc, const Allocation *alloc) {
|
|
if(!alloc->mHal.drvState.lod[0].mallocPtr) {
|
|
return;
|
|
}
|
|
uint32_t numFaces = alloc->getType()->getDimFaces() ? 6 : 1;
|
|
for (uint32_t face = 0; face < numFaces; face ++) {
|
|
for (uint32_t lod=0; lod < (alloc->getType()->getLODCount() -1); lod++) {
|
|
switch (alloc->getType()->getElement()->getSizeBits()) {
|
|
case 32:
|
|
mip8888(alloc, lod, (RsAllocationCubemapFace)face);
|
|
break;
|
|
case 16:
|
|
mip565(alloc, lod, (RsAllocationCubemapFace)face);
|
|
break;
|
|
case 8:
|
|
mip8(alloc, lod, (RsAllocationCubemapFace)face);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
uint32_t rsdAllocationGrallocBits(const android::renderscript::Context *rsc,
|
|
android::renderscript::Allocation *alloc)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
void rsdAllocationUpdateCachedObject(const Context *rsc,
|
|
const Allocation *alloc,
|
|
rs_allocation *obj)
|
|
{
|
|
obj->p = alloc;
|
|
#ifdef __LP64__
|
|
obj->unused1 = nullptr;
|
|
obj->unused2 = nullptr;
|
|
obj->unused3 = nullptr;
|
|
#endif
|
|
}
|