/* * Copyright (c) 2011-2020, The Linux Foundation. All rights reserved. * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials provided * with the distribution. * * Neither the name of The Linux Foundation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED "AS IS" AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include #include "gr_adreno_info.h" #include "gr_utils.h" #define ASTC_BLOCK_SIZE 16 #ifndef COLOR_FMT_P010_UBWC #define COLOR_FMT_P010_UBWC 9 #endif namespace gralloc { bool IsYuvFormat(int format) { switch (format) { case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: // Same as YCbCr_420_SP_VENUS case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV21_ZSL: case HAL_PIXEL_FORMAT_RAW16: case HAL_PIXEL_FORMAT_Y16: case HAL_PIXEL_FORMAT_RAW12: case HAL_PIXEL_FORMAT_RAW10: case HAL_PIXEL_FORMAT_YV12: case HAL_PIXEL_FORMAT_Y8: case HAL_PIXEL_FORMAT_YCbCr_420_P010: case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: // Below formats used by camera and VR case HAL_PIXEL_FORMAT_BLOB: case HAL_PIXEL_FORMAT_RAW_OPAQUE: case HAL_PIXEL_FORMAT_NV12_HEIF: case HAL_PIXEL_FORMAT_CbYCrY_422_I: return true; default: return false; } } bool IsUncompressedRGBFormat(int format) { switch (format) { case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_RGB_888: case HAL_PIXEL_FORMAT_RGB_565: case HAL_PIXEL_FORMAT_BGR_565: case HAL_PIXEL_FORMAT_BGRA_8888: case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: case HAL_PIXEL_FORMAT_R_8: case HAL_PIXEL_FORMAT_RG_88: case HAL_PIXEL_FORMAT_BGRX_8888: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_ARGB_2101010: case HAL_PIXEL_FORMAT_RGBX_1010102: case HAL_PIXEL_FORMAT_XRGB_2101010: case HAL_PIXEL_FORMAT_BGRA_1010102: case HAL_PIXEL_FORMAT_ABGR_2101010: case HAL_PIXEL_FORMAT_BGRX_1010102: case HAL_PIXEL_FORMAT_XBGR_2101010: case HAL_PIXEL_FORMAT_RGBA_FP16: case HAL_PIXEL_FORMAT_BGR_888: return true; default: break; } return false; } bool IsCompressedRGBFormat(int format) { switch (format) { case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_4x4_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_4x4_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_5x4_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_5x4_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_5x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_5x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_6x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_6x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_6x6_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_6x6_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_8x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_8x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_8x6_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_8x6_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_8x8_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_8x8_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x5_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x6_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x6_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x8_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x8_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x10_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x10_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_12x10_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_12x10_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_12x12_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_12x12_KHR: return true; default: break; } return false; } uint32_t GetBppForUncompressedRGB(int format) { uint32_t bpp = 0; switch (format) { case HAL_PIXEL_FORMAT_RGBA_FP16: bpp = 8; break; case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_BGRA_8888: case HAL_PIXEL_FORMAT_BGRX_8888: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_ARGB_2101010: case HAL_PIXEL_FORMAT_RGBX_1010102: case HAL_PIXEL_FORMAT_XRGB_2101010: case HAL_PIXEL_FORMAT_BGRA_1010102: case HAL_PIXEL_FORMAT_ABGR_2101010: case HAL_PIXEL_FORMAT_BGRX_1010102: case HAL_PIXEL_FORMAT_XBGR_2101010: bpp = 4; break; case HAL_PIXEL_FORMAT_RGB_888: case HAL_PIXEL_FORMAT_BGR_888: bpp = 3; break; case HAL_PIXEL_FORMAT_RGB_565: case HAL_PIXEL_FORMAT_BGR_565: case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: case HAL_PIXEL_FORMAT_RG_88: bpp = 2; break; case HAL_PIXEL_FORMAT_R_8: bpp = 1; break; default: ALOGE("Error : %s New format request = 0x%x", __FUNCTION__, format); break; } return bpp; } bool CpuCanAccess(uint64_t usage) { return CpuCanRead(usage) || CpuCanWrite(usage); } bool CpuCanRead(uint64_t usage) { if (usage & BufferUsage::CPU_READ_MASK) { return true; } return false; } bool CpuCanWrite(uint64_t usage) { if (usage & BufferUsage::CPU_WRITE_MASK) { // Application intends to use CPU for rendering return true; } return false; } uint32_t GetDataAlignment(int format, uint64_t usage) { uint32_t align = UINT(getpagesize()); if (format == HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED) { align = SIZE_8K; } if (usage & BufferUsage::PROTECTED) { if (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY) { align = SZ_2M; } else { align = SECURE_ALIGN; } } return align; } bool IsGPUFlagSupported(uint64_t usage) { bool ret = true; if ((usage & BufferUsage::GPU_MIPMAP_COMPLETE)) { ALOGE("GPU_MIPMAP_COMPLETE not supported"); ret = false; } if ((usage & BufferUsage::GPU_CUBE_MAP)) { ALOGE("GPU_CUBE_MAP not supported"); ret = false; } return ret; } int GetBpp(int format) { if (IsUncompressedRGBFormat(format)) { return GetBppForUncompressedRGB(format); } switch (format) { case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_4x4_KHR: case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_4x4_KHR: case HAL_PIXEL_FORMAT_RAW8: case HAL_PIXEL_FORMAT_Y8: return 1; case HAL_PIXEL_FORMAT_RAW16: case HAL_PIXEL_FORMAT_Y16: case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_YCbCr_422_I: case HAL_PIXEL_FORMAT_YCrCb_422_I: case HAL_PIXEL_FORMAT_CbYCrY_422_I: return 2; case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: case HAL_PIXEL_FORMAT_YCbCr_420_P010: return 3; default: return -1; } } // Returns the final buffer size meant to be allocated with ion unsigned int GetSize(const BufferInfo &info, unsigned int alignedw, unsigned int alignedh) { unsigned int size = 0; int format = info.format; int width = info.width; int height = info.height; uint64_t usage = info.usage; if (!IsGPUFlagSupported(usage)) { ALOGE("Unsupported GPU usage flags present 0x%" PRIx64, usage); return 0; } if (IsUBwcEnabled(format, usage)) { size = GetUBwcSize(width, height, format, alignedw, alignedh); } else if (IsUncompressedRGBFormat(format)) { uint32_t bpp = GetBppForUncompressedRGB(format); size = alignedw * alignedh * bpp; } else if (IsCompressedRGBFormat(format)) { size = alignedw * alignedh * ASTC_BLOCK_SIZE; } else { // Below switch should be for only YUV/custom formats switch (format) { case HAL_PIXEL_FORMAT_RAW16: case HAL_PIXEL_FORMAT_Y16:size = alignedw * alignedh * 2; break; case HAL_PIXEL_FORMAT_RAW10: case HAL_PIXEL_FORMAT_RAW12:size = ALIGN(alignedw * alignedh, SIZE_4K); break; case HAL_PIXEL_FORMAT_RAW8: case HAL_PIXEL_FORMAT_Y8:size = alignedw * alignedh * 1; break; // adreno formats case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: // NV21 size = ALIGN(alignedw * alignedh, SIZE_4K); size += (unsigned int) ALIGN(2 * ALIGN(width / 2, 32) * ALIGN(height / 2, 32), SIZE_4K); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: // NV12 // The chroma plane is subsampled, // but the pitch in bytes is unchanged // The GPU needs 4K alignment, but the video decoder needs 8K size = ALIGN(alignedw * alignedh, SIZE_8K); size += ALIGN(alignedw * (unsigned int) ALIGN(height / 2, 32), SIZE_8K); break; case HAL_PIXEL_FORMAT_YV12: if ((format == HAL_PIXEL_FORMAT_YV12) && ((width & 1) || (height & 1))) { ALOGE("w or h is odd for the YV12 format"); return 0; } size = alignedw * alignedh + (ALIGN(alignedw / 2, 16) * (alignedh / 2)) * 2; size = ALIGN(size, (unsigned int) SIZE_4K); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP: size = ALIGN((alignedw * alignedh) + (alignedw * alignedh) / 2 + 1, SIZE_4K); break; case HAL_PIXEL_FORMAT_YCbCr_420_P010: size = ALIGN((alignedw * alignedh * 2) + (alignedw * alignedh) + 1, SIZE_4K); break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: size = VENUS_BUFFER_SIZE(COLOR_FMT_P010, width, height); break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_YCbCr_422_I: case HAL_PIXEL_FORMAT_YCrCb_422_I: case HAL_PIXEL_FORMAT_CbYCrY_422_I: if (width & 1) { ALOGE("width is odd for the YUV422_SP format"); return 0; } size = ALIGN(alignedw * alignedh * 2, SIZE_4K); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12, width, height); break; case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: size = VENUS_BUFFER_SIZE(COLOR_FMT_NV21, width, height); break; case HAL_PIXEL_FORMAT_BLOB: case HAL_PIXEL_FORMAT_RAW_OPAQUE: if (height != 1) { ALOGE("%s: Buffers with HAL_PIXEL_FORMAT_BLOB must have height 1 ", __FUNCTION__); return 0; } size = (unsigned int) width; break; case HAL_PIXEL_FORMAT_NV21_ZSL: size = ALIGN((alignedw * alignedh) + (alignedw * alignedh) / 2, SIZE_4K); break; case HAL_PIXEL_FORMAT_NV12_HEIF: size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_512, width, height); break; default:ALOGE("%s: Unrecognized pixel format: 0x%x", __FUNCTION__, format); return 0; } } auto align = GetDataAlignment(format, usage); size = ALIGN(size, align) * info.layer_count; return size; } int GetBufferSizeAndDimensions(const BufferInfo &info, unsigned int *size, unsigned int *alignedw, unsigned int *alignedh) { GraphicsMetadata graphics_metadata = {}; return GetBufferSizeAndDimensions(info, size, alignedw, alignedh, &graphics_metadata); } int GetBufferSizeAndDimensions(const BufferInfo &info, unsigned int *size, unsigned int *alignedw, unsigned int *alignedh, GraphicsMetadata *graphics_metadata) { int buffer_type = GetBufferType(info.format); if (CanUseAdrenoForSize(buffer_type, info.usage)) { return GetGpuResourceSizeAndDimensions(info, size, alignedw, alignedh, graphics_metadata); } else { GetAlignedWidthAndHeight(info, alignedw, alignedh); *size = GetSize(info, *alignedw, *alignedh); } return 0; } void GetYuvUbwcSPPlaneInfo(uint32_t width, uint32_t height, int color_format, PlaneLayoutInfo *plane_info) { // UBWC buffer has these 4 planes in the following sequence: // Y_Plane, UV_Plane, Y_Meta_Plane, UV_Meta_Plane unsigned int y_meta_stride, y_meta_height, y_meta_size; unsigned int y_stride, y_height, y_size; unsigned int c_meta_stride, c_meta_height, c_meta_size; unsigned int alignment = 4096; unsigned int c_stride, c_height, c_size; uint64_t yOffset, cOffset, yMetaOffset, cMetaOffset; y_meta_stride = VENUS_Y_META_STRIDE(color_format, INT(width)); y_meta_height = VENUS_Y_META_SCANLINES(color_format, INT(height)); y_meta_size = ALIGN((y_meta_stride * y_meta_height), alignment); y_stride = VENUS_Y_STRIDE(color_format, INT(width)); y_height = VENUS_Y_SCANLINES(color_format, INT(height)); y_size = ALIGN((y_stride * y_height), alignment); c_meta_stride = VENUS_UV_META_STRIDE(color_format, INT(width)); c_meta_height = VENUS_UV_META_SCANLINES(color_format, INT(height)); c_meta_size = ALIGN((c_meta_stride * c_meta_height), alignment); c_stride = VENUS_UV_STRIDE(color_format, INT(width)); c_height = VENUS_UV_SCANLINES(color_format, INT(height)); c_size = ALIGN((c_stride * c_height), alignment); yMetaOffset = 0; yOffset = y_meta_size; cMetaOffset = y_meta_size + y_size; cOffset = y_meta_size + y_size + c_meta_size; plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; plane_info[0].offset = (uint32_t)yOffset; plane_info[0].stride = static_cast(UINT(width)); plane_info[0].stride_bytes = static_cast(y_stride); plane_info[0].scanlines = static_cast(y_height); plane_info[0].size = static_cast(y_size); plane_info[1].component = (PlaneComponent)(PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); plane_info[1].offset = (uint32_t)cOffset; plane_info[1].stride = static_cast(UINT(width)); plane_info[1].stride_bytes = static_cast(c_stride); plane_info[1].scanlines = static_cast(c_height); plane_info[1].size = static_cast(c_size); plane_info[2].component = (PlaneComponent)(PLANE_COMPONENT_META | PLANE_COMPONENT_Y); plane_info[2].offset = (uint32_t)yMetaOffset; plane_info[2].stride = static_cast(UINT(width)); plane_info[2].stride_bytes = static_cast(y_meta_stride); plane_info[2].scanlines = static_cast(y_meta_height); plane_info[2].size = static_cast(y_meta_size); plane_info[3].component = (PlaneComponent)(PLANE_COMPONENT_META | PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); plane_info[3].offset = (uint32_t)cMetaOffset; plane_info[3].stride = static_cast(UINT(width)); plane_info[3].stride_bytes = static_cast(c_meta_stride); plane_info[3].scanlines = static_cast(c_meta_height); plane_info[3].size = static_cast(c_meta_size); } // This API gets information about 8 planes (Y_Plane, UV_Plane, Y_Meta_Plane, UV_Meta_Plane, // Y_Plane, UV_Plane, Y_Meta_Plane, UV_Meta_Plane) and it stores the // information in PlaneLayoutInfo array. void GetYuvUbwcInterlacedSPPlaneInfo(uint32_t width, uint32_t height, PlaneLayoutInfo plane_info[8]) { // UBWC interlaced has top-bottom field layout with each field as // 8-plane (including meta plane also) NV12_UBWC with width = image_width // & height = image_height / 2. // Client passed plane_info argument is ptr to struct PlaneLayoutInfo[8]. // Plane info to be filled for each field separately. height = (height + 1) >> 1; GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_UBWC, &plane_info[0]); GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_UBWC, &plane_info[4]); } // This API gets information about 2 planes (Y_Plane & UV_Plane) void GetYuvSPPlaneInfo(const BufferInfo &info, int format, uint32_t width, uint32_t height, uint32_t bpp, PlaneLayoutInfo *plane_info) { int unaligned_width = info.width; int unaligned_height = info.height; unsigned int y_stride = 0, y_height = 0, y_size = 0; unsigned int c_stride = 0, c_height = 0, c_size = 0; uint64_t yOffset, cOffset; y_stride = c_stride = UINT(width) * bpp; y_height = INT(height); y_size = y_stride * y_height; switch (format) { case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP: c_size = (width * height) / 2 + 1; c_height = height >> 1; break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: if (unaligned_width & 1) { ALOGE("width is odd for the YUV422_SP format"); return; } c_size = width * height; c_height = height; break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12, height); c_size = c_stride * c_height; break; case HAL_PIXEL_FORMAT_NV12_HEIF: c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12_512, height); c_size = c_stride * c_height; break; case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: y_size = ALIGN(width * height, 4096); c_size = ALIGN(2 * ALIGN(unaligned_width / 2, 32) * ALIGN(unaligned_height / 2, 32), 4096); break; case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV21, height); c_size = c_stride * c_height; break; case HAL_PIXEL_FORMAT_NV21_ZSL: c_size = (width * height) / 2; c_height = height >> 1; break; case HAL_PIXEL_FORMAT_Y16: c_size = width * height; c_height = height; break; case HAL_PIXEL_FORMAT_Y8: c_size = 0; break; case HAL_PIXEL_FORMAT_YCbCr_420_P010: c_size = (width * height) + 1; c_height = height; break; default: break; } yOffset = 0; cOffset = y_size; plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; plane_info[0].offset = (uint32_t)yOffset; plane_info[0].step = 1; plane_info[0].stride = static_cast(UINT(width)); plane_info[0].stride_bytes = static_cast(y_stride); plane_info[0].scanlines = static_cast(y_height); plane_info[0].size = static_cast(y_size); plane_info[1].component = (PlaneComponent)(PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); plane_info[1].offset = (uint32_t)cOffset; plane_info[1].step = 2 * bpp; plane_info[1].stride = static_cast(UINT(width)); plane_info[1].stride_bytes = static_cast(c_stride); plane_info[1].scanlines = static_cast(c_height); plane_info[1].size = static_cast(c_size); } int GetYUVPlaneInfo(const private_handle_t *hnd, struct android_ycbcr ycbcr[2]) { int err = 0; uint32_t width = UINT(hnd->width); uint32_t height = UINT(hnd->height); int format = hnd->format; uint64_t usage = hnd->usage; int32_t interlaced = 0; int plane_count = 0; int unaligned_width = INT(hnd->unaligned_width); int unaligned_height = INT(hnd->unaligned_height); BufferInfo info(unaligned_width, unaligned_height, format, usage); memset(ycbcr->reserved, 0, sizeof(ycbcr->reserved)); // Check if UBWC buffer has been rendered in linear format. int linear_format = 0; if (getMetaData(const_cast(hnd), GET_LINEAR_FORMAT, &linear_format) == 0) { format = INT(linear_format); } // Check metadata if the geometry has been updated. BufferDim_t buffer_dim; if (getMetaData(const_cast(hnd), GET_BUFFER_GEOMETRY, &buffer_dim) == 0) { BufferInfo info(buffer_dim.sliceWidth, buffer_dim.sliceHeight, format, usage); GetAlignedWidthAndHeight(info, &width, &height); } // Check metadata for interlaced content. int interlace_flag = 0; if (getMetaData(const_cast(hnd), GET_PP_PARAM_INTERLACED, &interlace_flag) == 0) { if (interlace_flag) { interlaced = LAYOUT_INTERLACED_FLAG; } } PlaneLayoutInfo plane_info[8] = {}; // Get the chroma offsets from the handle width/height. We take advantage // of the fact the width _is_ the stride err = GetYUVPlaneInfo(info, format, width, height, interlaced, &plane_count, plane_info); if (err == 0) { if (interlaced && format == HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC) { CopyPlaneLayoutInfotoAndroidYcbcr(hnd->base, plane_count, &plane_info[0], &ycbcr[0]); unsigned int uv_stride, uv_height, uv_size; unsigned int alignment = 4096; uint64_t field_base; height = (height + 1) >> 1; uv_stride = VENUS_UV_STRIDE(COLOR_FMT_NV12_UBWC, INT(width)); uv_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12_UBWC, INT(height)); uv_size = ALIGN((uv_stride * uv_height), alignment); field_base = hnd->base + plane_info[1].offset + uv_size; memset(ycbcr[1].reserved, 0, sizeof(ycbcr[1].reserved)); CopyPlaneLayoutInfotoAndroidYcbcr(field_base, plane_count, &plane_info[4], &ycbcr[1]); } else { CopyPlaneLayoutInfotoAndroidYcbcr(hnd->base, plane_count, plane_info, ycbcr); switch (format) { case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV21_ZSL: case HAL_PIXEL_FORMAT_RAW16: case HAL_PIXEL_FORMAT_Y16: case HAL_PIXEL_FORMAT_RAW10: case HAL_PIXEL_FORMAT_RAW8: case HAL_PIXEL_FORMAT_Y8: std::swap(ycbcr->cb, ycbcr->cr); } } } return err; } int GetRawPlaneInfo(int32_t format, int32_t width, int32_t height, PlaneLayoutInfo *plane_info) { int32_t step = 0; switch (format) { case HAL_PIXEL_FORMAT_RAW16: step = 2; break; case HAL_PIXEL_FORMAT_RAW8: step = 1; break; case HAL_PIXEL_FORMAT_RAW12: case HAL_PIXEL_FORMAT_RAW10: case HAL_PIXEL_FORMAT_BLOB: step = 0; break; default: ALOGW("RawPlaneInfo is unsupported for format 0x%x", format); return -EINVAL; } BufferInfo info(width, height, format); uint32_t alignedWidth, alignedHeight; GetAlignedWidthAndHeight(info, &alignedWidth, &alignedHeight); uint32_t size = GetSize(info, alignedWidth, alignedHeight); plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_RAW; plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[0].offset = 0; plane_info[0].step = step; plane_info[0].stride = width; plane_info[0].stride_bytes = static_cast(alignedWidth); if (format == HAL_PIXEL_FORMAT_RAW16) { plane_info[0].stride_bytes = static_cast(alignedWidth * GetBpp(format)); } plane_info[0].scanlines = height; plane_info[0].size = size; return 0; } // Explicitly defined UBWC formats bool IsUBwcFormat(int format) { switch (format) { case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: return true; default: return false; } } bool IsUBwcSupported(int format) { // Existing HAL formats with UBWC support switch (format) { case HAL_PIXEL_FORMAT_BGR_565: case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_RGBX_1010102: case HAL_PIXEL_FORMAT_DEPTH_16: case HAL_PIXEL_FORMAT_DEPTH_24: case HAL_PIXEL_FORMAT_DEPTH_24_STENCIL_8: case HAL_PIXEL_FORMAT_DEPTH_32F: case HAL_PIXEL_FORMAT_STENCIL_8: return true; default: break; } return false; } bool IsUBwcPISupported(int format, uint64_t usage) { if (usage & BufferUsage::COMPOSER_OVERLAY || !(usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC_PI)) { return false; } // As of now only two formats switch (format) { case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: { if ((usage & BufferUsage::GPU_TEXTURE) || (usage & BufferUsage::GPU_RENDER_TARGET)) { if (AdrenoMemInfo::GetInstance()) { return AdrenoMemInfo::GetInstance()->IsPISupportedByGPU(format, usage); } } else { return true; } } } return false; } bool IsUBwcEnabled(int format, uint64_t usage) { // Allow UBWC, if client is using an explicitly defined UBWC pixel format. if (IsUBwcFormat(format)) { return true; } // Allow UBWC, if an OpenGL client sets UBWC usage flag and GPU plus MDP // support the format. OR if a non-OpenGL client like Rotator, sets UBWC // usage flag and MDP supports the format. if (((usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC) || (usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC_PI) || (usage & BufferUsage::COMPOSER_CLIENT_TARGET)) && IsUBwcSupported(format)) { bool enable = true; // Query GPU for UBWC only if buffer is intended to be used by GPU. if ((usage & BufferUsage::GPU_TEXTURE) || (usage & BufferUsage::GPU_RENDER_TARGET)) { if (AdrenoMemInfo::GetInstance()) { enable = AdrenoMemInfo::GetInstance()->IsUBWCSupportedByGPU(format); } } // Allow UBWC, only if CPU usage flags are not set if (enable && !(CpuCanAccess(usage))) { return true; } } return false; } void GetYuvUBwcWidthAndHeight(int width, int height, int format, unsigned int *aligned_w, unsigned int *aligned_h) { switch (format) { case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: *aligned_w = VENUS_Y_STRIDE(COLOR_FMT_NV12, width); *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_NV12, height); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: *aligned_w = VENUS_Y_STRIDE(COLOR_FMT_NV12_UBWC, width); *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_NV12_UBWC, height); break; case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: // The macro returns the stride which is 4/3 times the width, hence * 3/4 *aligned_w = (VENUS_Y_STRIDE(COLOR_FMT_NV12_BPP10_UBWC, width) * 3) / 4; *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_NV12_BPP10_UBWC, height); break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: // The macro returns the stride which is 2 times the width, hence / 2 *aligned_w = (VENUS_Y_STRIDE(COLOR_FMT_P010_UBWC, width) / 2); *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_P010_UBWC, height); break; default: ALOGE("%s: Unsupported pixel format: 0x%x", __FUNCTION__, format); *aligned_w = 0; *aligned_h = 0; break; } } void GetRgbUBwcBlockSize(uint32_t bpp, int *block_width, int *block_height) { *block_width = 0; *block_height = 0; switch (bpp) { case 2: case 4: *block_width = 16; *block_height = 4; break; case 8: *block_width = 8; *block_height = 4; break; case 16: *block_width = 4; *block_height = 4; break; default: ALOGE("%s: Unsupported bpp: %d", __FUNCTION__, bpp); break; } } unsigned int GetRgbUBwcMetaBufferSize(int width, int height, uint32_t bpp) { unsigned int size = 0; int meta_width, meta_height; int block_width, block_height; GetRgbUBwcBlockSize(bpp, &block_width, &block_height); if (!block_width || !block_height) { ALOGE("%s: Unsupported bpp: %d", __FUNCTION__, bpp); return size; } // Align meta buffer height to 16 blocks meta_height = ALIGN(((height + block_height - 1) / block_height), 16); // Align meta buffer width to 64 blocks meta_width = ALIGN(((width + block_width - 1) / block_width), 64); // Align meta buffer size to 4K size = (unsigned int)ALIGN((meta_width * meta_height), 4096); return size; } unsigned int GetUBwcSize(int width, int height, int format, unsigned int alignedw, unsigned int alignedh) { unsigned int size = 0; uint32_t bpp = 0; switch (format) { case HAL_PIXEL_FORMAT_BGR_565: case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_RGBX_1010102: bpp = GetBppForUncompressedRGB(format); size = alignedw * alignedh * bpp; size += GetRgbUBwcMetaBufferSize(width, height, bpp); break; /* * 1. The CtsMediaV2TestCases#CodecEncoderSurfaceTest is a transcode use case and shares * same surface between encoder and decoder. * 2. Configures encoder with Opaque color format thus encoder sets ubwc usage bits and * is configured with NV12_UBWC format. * 3. Configures decoder as 'flexible', thus configuring decoder with NV12 format. * 4. Decoder should produce output to surface that will be fed back to encoder as input. * 5. Though UBWC is enabled, we need to compute the actual buffer size (including aligned * width and height) based on pixel format that is set. */ case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12, width, height); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_UBWC, width, height); break; case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_BPP10_UBWC, width, height); break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: size = VENUS_BUFFER_SIZE(COLOR_FMT_P010_UBWC, width, height); break; default: ALOGE("%s: Unsupported pixel format: 0x%x", __FUNCTION__, format); break; } return size; } unsigned int GetRgbMetaSize(int format, uint32_t width, uint32_t height, uint64_t usage) { unsigned int meta_size = 0; if (!IsUBwcEnabled(format, usage)) { return meta_size; } uint32_t bpp = GetBppForUncompressedRGB(format); switch (format) { case HAL_PIXEL_FORMAT_BGR_565: case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_RGBX_1010102: case HAL_PIXEL_FORMAT_RGBA_FP16: meta_size = GetRgbUBwcMetaBufferSize(width, height, bpp); break; default: ALOGE("%s:Unsupported RGB format: 0x%x", __FUNCTION__, format); break; } return meta_size; } int GetRgbDataAddress(private_handle_t *hnd, void **rgb_data) { int err = 0; // This api is for RGB* formats if (!IsUncompressedRGBFormat(hnd->format)) { return -EINVAL; } // linear buffer, nothing to do further if (!(hnd->flags & private_handle_t::PRIV_FLAGS_UBWC_ALIGNED)) { *rgb_data = reinterpret_cast(hnd->base); return err; } unsigned int meta_size = GetRgbMetaSize(hnd->format, hnd->width, hnd->height, hnd->usage); *rgb_data = reinterpret_cast(hnd->base + meta_size); return err; } void GetCustomDimensions(private_handle_t *hnd, int *stride, int *height) { BufferDim_t buffer_dim; int interlaced = 0; *stride = hnd->width; *height = hnd->height; if (getMetaData(hnd, GET_BUFFER_GEOMETRY, &buffer_dim) == 0) { *stride = buffer_dim.sliceWidth; *height = buffer_dim.sliceHeight; } else if (getMetaData(hnd, GET_PP_PARAM_INTERLACED, &interlaced) == 0) { if (interlaced && IsUBwcFormat(hnd->format)) { unsigned int alignedw = 0, alignedh = 0; // Get re-aligned height for single ubwc interlaced field and // multiply by 2 to get frame height. BufferInfo info(hnd->width, ((hnd->height + 1) >> 1), hnd->format); GetAlignedWidthAndHeight(info, &alignedw, &alignedh); *stride = static_cast(alignedw); *height = static_cast(alignedh * 2); } } } void GetColorSpaceFromMetadata(private_handle_t *hnd, int *color_space) { ColorMetaData color_metadata; if (getMetaData(hnd, GET_COLOR_METADATA, &color_metadata) == 0) { switch (color_metadata.colorPrimaries) { case ColorPrimaries_BT709_5: *color_space = HAL_CSC_ITU_R_709; break; case ColorPrimaries_BT601_6_525: case ColorPrimaries_BT601_6_625: *color_space = ((color_metadata.range) ? HAL_CSC_ITU_R_601_FR : HAL_CSC_ITU_R_601); break; case ColorPrimaries_BT2020: *color_space = (color_metadata.range) ? HAL_CSC_ITU_R_2020_FR : HAL_CSC_ITU_R_2020; break; default: ALOGE("Unknown Color Space = %d", color_metadata.colorPrimaries); break; } } else if (getMetaData(hnd, GET_COLOR_SPACE, color_space) != 0) { *color_space = 0; } } void GetAlignedWidthAndHeight(const BufferInfo &info, unsigned int *alignedw, unsigned int *alignedh) { int width = info.width; int height = info.height; int format = info.format; uint64_t usage = info.usage; // Currently surface padding is only computed for RGB* surfaces. bool ubwc_enabled = IsUBwcEnabled(format, usage); int tile = ubwc_enabled; if (IsUncompressedRGBFormat(format)) { if (AdrenoMemInfo::GetInstance()) { AdrenoMemInfo::GetInstance()->AlignUnCompressedRGB(width, height, format, tile, alignedw, alignedh); } return; } if (ubwc_enabled) { GetYuvUBwcWidthAndHeight(width, height, format, alignedw, alignedh); return; } if (IsCompressedRGBFormat(format)) { if (AdrenoMemInfo::GetInstance()) { AdrenoMemInfo::GetInstance()->AlignCompressedRGB(width, height, format, alignedw, alignedh); } return; } int aligned_w = width; int aligned_h = height; unsigned int alignment = 32; // Below should be only YUV family switch (format) { case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YCbCr_420_SP: if (AdrenoMemInfo::GetInstance() == nullptr) { return; } alignment = AdrenoMemInfo::GetInstance()->GetGpuPixelAlignment(); aligned_w = ALIGN(width, alignment); break; case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: aligned_w = ALIGN(width, alignment); break; case HAL_PIXEL_FORMAT_RAW16: case HAL_PIXEL_FORMAT_Y16: case HAL_PIXEL_FORMAT_Y8: aligned_w = ALIGN(width, 16); break; case HAL_PIXEL_FORMAT_RAW12: aligned_w = ALIGN(width * 12 / 8, 16); break; case HAL_PIXEL_FORMAT_RAW10: { const unsigned int gpu_alignment = AdrenoMemInfo::GetInstance()->GetGpuPixelAlignment(); // gpu_alignment can return 1. Make sure it's at least 64. const unsigned int raw10_alignment = std::max(gpu_alignment, 64u); aligned_w = ALIGN(width * 10 / 8, raw10_alignment); } break; case HAL_PIXEL_FORMAT_RAW8: aligned_w = ALIGN(width, 16); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: aligned_w = ALIGN(width, 128); break; case HAL_PIXEL_FORMAT_YV12: if ((usage & BufferUsage::GPU_TEXTURE) || (usage & BufferUsage::GPU_RENDER_TARGET)) { if (AdrenoMemInfo::GetInstance() == nullptr) { return; } alignment = AdrenoMemInfo::GetInstance()->GetGpuPixelAlignment(); aligned_w = ALIGN(width, alignment); } else { aligned_w = ALIGN(width, 16); } break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_YCbCr_422_I: case HAL_PIXEL_FORMAT_YCrCb_422_I: case HAL_PIXEL_FORMAT_YCbCr_420_P010: aligned_w = ALIGN(width, 16); break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_P010, width) / 2); aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_P010, height)); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV12, width)); aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV12, height)); break; case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV21, width)); aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV21, height)); break; case HAL_PIXEL_FORMAT_BLOB: case HAL_PIXEL_FORMAT_RAW_OPAQUE: break; case HAL_PIXEL_FORMAT_NV21_ZSL: aligned_w = ALIGN(width, 64); aligned_h = ALIGN(height, 64); break; case HAL_PIXEL_FORMAT_NV12_HEIF: aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV12_512, width)); aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV12_512, height)); break; default: break; } *alignedw = (unsigned int)aligned_w; *alignedh = (unsigned int)aligned_h; } int GetBufferLayout(private_handle_t *hnd, uint32_t stride[4], uint32_t offset[4], uint32_t *num_planes) { if (!hnd || !stride || !offset || !num_planes) { return -EINVAL; } struct android_ycbcr yuvPlaneInfo[2] = {}; *num_planes = 1; stride[0] = 0; switch (hnd->format) { case HAL_PIXEL_FORMAT_RGB_565: case HAL_PIXEL_FORMAT_BGR_565: case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: stride[0] = static_cast(hnd->width * 2); break; case HAL_PIXEL_FORMAT_RGB_888: stride[0] = static_cast(hnd->width * 3); break; case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_BGRA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_BGRX_8888: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_ARGB_2101010: case HAL_PIXEL_FORMAT_RGBX_1010102: case HAL_PIXEL_FORMAT_XRGB_2101010: case HAL_PIXEL_FORMAT_BGRA_1010102: case HAL_PIXEL_FORMAT_ABGR_2101010: case HAL_PIXEL_FORMAT_BGRX_1010102: case HAL_PIXEL_FORMAT_XBGR_2101010: stride[0] = static_cast(hnd->width * 4); break; } // Format is RGB if (stride[0]) { return 0; } (*num_planes)++; int ret = GetYUVPlaneInfo(hnd, yuvPlaneInfo); if (ret < 0) { ALOGE("%s failed", __FUNCTION__); return ret; } // We are only returning buffer layout for progressive or single field formats. struct android_ycbcr yuvInfo = yuvPlaneInfo[0]; stride[0] = static_cast(yuvInfo.ystride); offset[0] = static_cast(reinterpret_cast(yuvInfo.y) - hnd->base); stride[1] = static_cast(yuvInfo.cstride); switch (hnd->format) { case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010: case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: case HAL_PIXEL_FORMAT_NV12_HEIF: offset[1] = static_cast(reinterpret_cast(yuvInfo.cb) - hnd->base); break; case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: offset[1] = static_cast(reinterpret_cast(yuvInfo.cr) - hnd->base); break; case HAL_PIXEL_FORMAT_YV12: offset[1] = static_cast(reinterpret_cast(yuvInfo.cr) - hnd->base); stride[2] = static_cast(yuvInfo.cstride); offset[2] = static_cast(reinterpret_cast(yuvInfo.cb) - hnd->base); (*num_planes)++; break; case HAL_PIXEL_FORMAT_CbYCrY_422_I: *num_planes = 1; break; default: ALOGW("%s: Unsupported format", __FUNCTION__); ret = -EINVAL; } if (hnd->flags & private_handle_t::PRIV_FLAGS_UBWC_ALIGNED) { std::fill(offset, offset + 4, 0); } return 0; } int GetGpuResourceSizeAndDimensions(const BufferInfo &info, unsigned int *size, unsigned int *alignedw, unsigned int *alignedh, GraphicsMetadata *graphics_metadata) { GetAlignedWidthAndHeight(info, alignedw, alignedh); AdrenoMemInfo* adreno_mem_info = AdrenoMemInfo::GetInstance(); graphics_metadata->size = adreno_mem_info->AdrenoGetMetadataBlobSize(); uint64_t adreno_usage = info.usage; // If gralloc disables UBWC based on any of the checks, // we pass modified usage flag to adreno to convey this. int is_ubwc_enabled = IsUBwcEnabled(info.format, info.usage); if (!is_ubwc_enabled) { adreno_usage &= ~(GRALLOC_USAGE_PRIVATE_ALLOC_UBWC); } else { adreno_usage |= GRALLOC_USAGE_PRIVATE_ALLOC_UBWC; } // Call adreno api for populating metadata blob // Layer count is for 2D/Cubemap arrays and depth is used for 3D slice // Using depth to pass layer_count here int ret = adreno_mem_info->AdrenoInitMemoryLayout(graphics_metadata->data, info.width, info.height, info.layer_count, /* depth */ info.format, 1, is_ubwc_enabled, adreno_usage, 1); if (ret != 0) { ALOGE("%s Graphics metadata init failed", __FUNCTION__); *size = 0; return -EINVAL; } // Call adreno api with the metadata blob to get buffer size *size = adreno_mem_info->AdrenoGetAlignedGpuBufferSize(graphics_metadata->data); return 0; } bool CanUseAdrenoForSize(int buffer_type, uint64_t usage) { if (buffer_type == BUFFER_TYPE_VIDEO || !GetAdrenoSizeAPIStatus()) { return false; } if ((usage & BufferUsage::PROTECTED) && ((usage & BufferUsage::CAMERA_OUTPUT) || (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY))) { return false; } return true; } bool GetAdrenoSizeAPIStatus() { AdrenoMemInfo* adreno_mem_info = AdrenoMemInfo::GetInstance(); if (adreno_mem_info) { return adreno_mem_info->AdrenoSizeAPIAvaliable(); } return false; } bool UseUncached(int format, uint64_t usage) { if ((usage & GRALLOC_USAGE_PRIVATE_UNCACHED) || (usage & BufferUsage::PROTECTED)) { return true; } // CPU read rarely if ((usage & BufferUsage::CPU_READ_MASK) == static_cast(BufferUsage::CPU_READ_RARELY)) { return true; } // CPU write rarely if ((usage & BufferUsage::CPU_WRITE_MASK) == static_cast(BufferUsage::CPU_WRITE_RARELY)) { return true; } if ((usage & BufferUsage::SENSOR_DIRECT_DATA) || (usage & BufferUsage::GPU_DATA_BUFFER)) { return true; } if (format && IsUBwcEnabled(format, usage)) { return true; } return false; } uint64_t GetHandleFlags(int format, uint64_t usage) { uint64_t priv_flags = 0; if (usage & BufferUsage::VIDEO_ENCODER) { priv_flags |= private_handle_t::PRIV_FLAGS_VIDEO_ENCODER; } if (usage & BufferUsage::CAMERA_OUTPUT) { priv_flags |= private_handle_t::PRIV_FLAGS_CAMERA_WRITE; } if (usage & BufferUsage::CAMERA_INPUT) { priv_flags |= private_handle_t::PRIV_FLAGS_CAMERA_READ; } if (usage & BufferUsage::COMPOSER_OVERLAY) { priv_flags |= private_handle_t::PRIV_FLAGS_DISP_CONSUMER; } if (usage & BufferUsage::GPU_TEXTURE) { priv_flags |= private_handle_t::PRIV_FLAGS_HW_TEXTURE; } if (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY) { priv_flags |= private_handle_t::PRIV_FLAGS_SECURE_DISPLAY; } if (IsUBwcEnabled(format, usage)) { if (IsUBwcPISupported(format, usage)) { priv_flags |= private_handle_t::PRIV_FLAGS_UBWC_ALIGNED_PI; } else { priv_flags |= private_handle_t::PRIV_FLAGS_UBWC_ALIGNED; } } if (usage & (BufferUsage::CPU_READ_MASK | BufferUsage::CPU_WRITE_MASK)) { priv_flags |= private_handle_t::PRIV_FLAGS_CPU_RENDERED; } if ((usage & (BufferUsage::VIDEO_ENCODER | BufferUsage::VIDEO_DECODER | BufferUsage::CAMERA_OUTPUT | BufferUsage::GPU_RENDER_TARGET))) { priv_flags |= private_handle_t::PRIV_FLAGS_NON_CPU_WRITER; } if (!UseUncached(format, usage)) { priv_flags |= private_handle_t::PRIV_FLAGS_CACHED; } return priv_flags; } int GetImplDefinedFormat(uint64_t usage, int format) { int gr_format = format; // If input format is HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED then based on // the usage bits, gralloc assigns a format. if (format == HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED || format == HAL_PIXEL_FORMAT_YCbCr_420_888) { if ((usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC || usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC_PI) && format != HAL_PIXEL_FORMAT_YCbCr_420_888) { gr_format = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC; } else if (usage & BufferUsage::VIDEO_ENCODER) { if (usage & GRALLOC_USAGE_PRIVATE_VIDEO_NV21_ENCODER) { gr_format = HAL_PIXEL_FORMAT_NV21_ENCODEABLE; // NV21 } else if (usage & GRALLOC_USAGE_PRIVATE_HEIF) { gr_format = HAL_PIXEL_FORMAT_NV12_HEIF; } else if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { gr_format = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS; } else { gr_format = HAL_PIXEL_FORMAT_NV12_ENCODEABLE; // NV12 } } else if (usage & BufferUsage::CAMERA_INPUT) { if (usage & BufferUsage::CAMERA_OUTPUT) { // Assumed ZSL if both producer and consumer camera flags set gr_format = HAL_PIXEL_FORMAT_NV21_ZSL; // NV21 } else { gr_format = HAL_PIXEL_FORMAT_YCrCb_420_SP; // NV21 } } else if (usage & BufferUsage::CAMERA_OUTPUT) { if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { gr_format = HAL_PIXEL_FORMAT_NV21_ZSL; // NV21 } else { gr_format = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS; // NV12 preview } } else if (usage & BufferUsage::COMPOSER_OVERLAY) { // XXX: If we still haven't set a format, default to RGBA8888 gr_format = HAL_PIXEL_FORMAT_RGBA_8888; } else if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { // If no other usage flags are detected, default the // flexible YUV format to NV21_ZSL gr_format = HAL_PIXEL_FORMAT_NV21_ZSL; } } return gr_format; } int GetCustomFormatFlags(int format, uint64_t usage, int *custom_format, uint64_t *priv_flags) { *custom_format = GetImplDefinedFormat(usage, format); *priv_flags = GetHandleFlags(*custom_format, usage); if (usage & GRALLOC_USAGE_PROTECTED) { *priv_flags |= private_handle_t::PRIV_FLAGS_SECURE_BUFFER; } *priv_flags |= private_handle_t::PRIV_FLAGS_USES_ION; return 0; } int GetBufferType(int inputFormat) { return IsYuvFormat(inputFormat) ? BUFFER_TYPE_VIDEO : BUFFER_TYPE_UI; } int GetYUVPlaneInfo(const BufferInfo &info, int32_t format, int32_t width, int32_t height, int32_t flags, int *plane_count, PlaneLayoutInfo *plane_info) { int err = 0; unsigned int y_stride, c_stride, y_height, c_height, y_size, c_size; uint64_t yOffset, cOffset, crOffset, cbOffset; int h_subsampling = 0, v_subsampling = 0; switch (format) { // Semiplanar case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: case HAL_PIXEL_FORMAT_NV12_HEIF: // Same as YCbCr_420_SP_VENUS case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_NV21_ZSL: case HAL_PIXEL_FORMAT_Y16: case HAL_PIXEL_FORMAT_Y8: *plane_count = 2; GetYuvSPPlaneInfo(info, format, width, height, 1, plane_info); GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[1].h_subsampling = h_subsampling; plane_info[1].v_subsampling = v_subsampling; break; case HAL_PIXEL_FORMAT_RAW10: case HAL_PIXEL_FORMAT_RAW8: case HAL_PIXEL_FORMAT_RAW16: case HAL_PIXEL_FORMAT_RAW12: case HAL_PIXEL_FORMAT_BLOB: *plane_count = 1; GetRawPlaneInfo(format, info.width, info.height, plane_info); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); if (flags & LAYOUT_INTERLACED_FLAG) { *plane_count = 8; GetYuvUbwcInterlacedSPPlaneInfo(width, height, plane_info); plane_info[0].step = plane_info[4].step = 1; plane_info[1].step = plane_info[5].step = 2; plane_info[0].h_subsampling = plane_info[4].h_subsampling = 0; plane_info[0].v_subsampling = plane_info[4].v_subsampling = 0; plane_info[1].h_subsampling = plane_info[5].h_subsampling = h_subsampling; plane_info[1].v_subsampling = plane_info[5].v_subsampling = v_subsampling; plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; plane_info[2].step = plane_info[3].step = 0; plane_info[6].h_subsampling = plane_info[7].h_subsampling = 0; plane_info[6].v_subsampling = plane_info[7].v_subsampling = 0; plane_info[6].step = plane_info[7].step = 0; } else { *plane_count = 4; GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_UBWC, plane_info); plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[0].step = 1; plane_info[1].h_subsampling = h_subsampling; plane_info[1].v_subsampling = v_subsampling; plane_info[1].step = 2; plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; plane_info[2].step = plane_info[3].step = 0; } break; case HAL_PIXEL_FORMAT_YCbCr_420_P010: *plane_count = 2; GetYuvSPPlaneInfo(info, format, width, height, 2, plane_info); GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[1].h_subsampling = h_subsampling; plane_info[1].v_subsampling = v_subsampling; break; case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: *plane_count = 4; GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_BPP10_UBWC, plane_info); GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[1].step = 1; plane_info[1].h_subsampling = h_subsampling; plane_info[1].v_subsampling = v_subsampling; plane_info[1].step = 3; plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; plane_info[2].step = plane_info[3].step = 0; break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: *plane_count = 4; GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_P010_UBWC, plane_info); GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[1].step = 1; plane_info[1].h_subsampling = h_subsampling; plane_info[1].v_subsampling = v_subsampling; plane_info[1].step = 4; plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; plane_info[2].step = plane_info[3].step = 0; break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: *plane_count = 2; y_stride = VENUS_Y_STRIDE(COLOR_FMT_P010, width); c_stride = VENUS_UV_STRIDE(COLOR_FMT_P010, width); y_height = VENUS_Y_SCANLINES(COLOR_FMT_P010, height); y_size = y_stride * y_height; yOffset = 0; cOffset = y_size; c_height = VENUS_UV_SCANLINES(COLOR_FMT_P010, INT(height)); c_size = c_stride * c_height; GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; plane_info[0].offset = (uint32_t)yOffset; plane_info[0].stride = static_cast(UINT(width)); plane_info[0].stride_bytes = static_cast(y_stride); plane_info[0].scanlines = static_cast(y_height); plane_info[0].size = static_cast(y_size); plane_info[0].step = 1; plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[1].component = (PlaneComponent)(PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); plane_info[1].offset = (uint32_t)cOffset; plane_info[1].stride = static_cast(UINT(width)); plane_info[1].stride_bytes = static_cast(c_stride); plane_info[1].scanlines = static_cast(c_height); plane_info[1].size = static_cast(c_size); plane_info[1].step = 4; plane_info[1].h_subsampling = h_subsampling; plane_info[1].v_subsampling = v_subsampling; break; // Planar case HAL_PIXEL_FORMAT_YV12: if ((info.width & 1) || (info.height & 1)) { ALOGE("w or h is odd for the YV12 format"); err = -EINVAL; return err; } *plane_count = 3; y_stride = width; c_stride = ALIGN(width / 2, 16); y_height = UINT(height); y_size = (y_stride * y_height); height = height >> 1; c_height = UINT(height); c_size = (c_stride * c_height); yOffset = 0; crOffset = y_size; cbOffset = (y_size + c_size); GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; plane_info[0].offset = (uint32_t)yOffset; plane_info[0].stride = static_cast(UINT(width)); plane_info[0].stride_bytes = static_cast(y_stride); plane_info[0].scanlines = static_cast(y_height); plane_info[0].size = static_cast(y_size); plane_info[0].step = 1; plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; plane_info[1].component = (PlaneComponent)PLANE_COMPONENT_Cb; plane_info[1].offset = (uint32_t)cbOffset; plane_info[2].component = (PlaneComponent)PLANE_COMPONENT_Cr; plane_info[2].offset = (uint32_t)crOffset; for (int i = 1; i < 3; i++) { plane_info[i].stride = static_cast(UINT(width)); plane_info[i].stride_bytes = static_cast(c_stride); plane_info[i].scanlines = static_cast(c_height); plane_info[i].size = static_cast(c_size); plane_info[i].step = 1; plane_info[i].h_subsampling = h_subsampling; plane_info[i].v_subsampling = v_subsampling; } break; case HAL_PIXEL_FORMAT_CbYCrY_422_I: if (info.width & 1) { ALOGE("width is odd for the YUV422_SP format"); err = -EINVAL; return err; } *plane_count = 1; y_stride = width * 2; y_height = UINT(height); y_size = y_stride * y_height; yOffset = 0; plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; plane_info[0].offset = (uint32_t)yOffset; plane_info[0].stride = static_cast(UINT(width)); plane_info[0].stride_bytes = static_cast(y_stride); plane_info[0].scanlines = static_cast(y_height); plane_info[0].size = static_cast(y_size); plane_info[0].step = 1; plane_info[0].h_subsampling = 0; plane_info[0].v_subsampling = 0; break; // Unsupported formats case HAL_PIXEL_FORMAT_YCbCr_422_I: case HAL_PIXEL_FORMAT_YCrCb_422_I: case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: default: *plane_count = 0; ALOGD("%s: Invalid format passed: 0x%x", __FUNCTION__, format); err = -EINVAL; } return err; } void GetYuvSubSamplingFactor(int32_t format, int *h_subsampling, int *v_subsampling) { switch (format) { case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCbCr_420_P010: case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: // Same as YCbCr_420_SP_VENUS case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: case HAL_PIXEL_FORMAT_NV21_ZSL: case HAL_PIXEL_FORMAT_YV12: *h_subsampling = 1; *v_subsampling = 1; break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: case HAL_PIXEL_FORMAT_CbYCrY_422_I: *h_subsampling = 1; *v_subsampling = 0; break; case HAL_PIXEL_FORMAT_Y16: case HAL_PIXEL_FORMAT_Y8: case HAL_PIXEL_FORMAT_BLOB: case HAL_PIXEL_FORMAT_RAW_OPAQUE: case HAL_PIXEL_FORMAT_NV12_HEIF: default: *h_subsampling = 0; *v_subsampling = 0; break; } } void CopyPlaneLayoutInfotoAndroidYcbcr(uint64_t base, int plane_count, PlaneLayoutInfo *plane_info, struct android_ycbcr *ycbcr) { ycbcr->y = reinterpret_cast(base + plane_info[0].offset); ycbcr->ystride = plane_info[0].stride_bytes; if (plane_count == 1) { ycbcr->cb = NULL; ycbcr->cr = NULL; ycbcr->cstride = 0; ycbcr->chroma_step = 0; } else if (plane_count == 2 || plane_count == 4 || plane_count == 8) { /* For YUV semiplanar :- * - In progressive & linear case plane count is 2 and plane_info[0] will * contain info about Y plane and plane_info[1] will contain info about UV plane. * - In progressive & compressed case plane count is 4 then plane_info[0] will * contain info about Y plane and plane_info[1] will contain info about UV plane. * Remaining two plane (plane_info[2] & plane_info[3]) contain info about the * Y_Meta_Plane and UV_Meta_Plane. * - In interlaced & compressed case plane count is 8 then plane_info[0], plane_info[1], * plane_info[4] & plane_info[5] will contain info about Y_plane, UV_plane, Y_plane * & UV_plane. Remaining plane will contain info about the meta planes. As in this case * this API is called twice through GetYUVPlaneInfo() with address of plane_info[0] & * plane_info[4], so this will calculate the information accordingly and will fill the * ycbcr structure with interlaced plane info only. */ ycbcr->cb = reinterpret_cast(base + plane_info[1].offset); ycbcr->cr = reinterpret_cast(base + plane_info[1].offset + 1); ycbcr->cstride = plane_info[1].stride_bytes; ycbcr->chroma_step = plane_info[1].step; } else if (plane_count == 3) { /* For YUV planar :- * Plane size is 3 and plane_info[0], plane_info[1], plane_info[2] will * contain info about y_plane, cb_plane and cr_plane accordingly. */ ycbcr->cb = reinterpret_cast(base + plane_info[1].offset); ycbcr->cr = reinterpret_cast(base + plane_info[2].offset); ycbcr->cstride = plane_info[1].stride_bytes; ycbcr->chroma_step = plane_info[1].step; } } bool HasAlphaComponent(int32_t format) { switch (format) { case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_BGRA_8888: case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: case HAL_PIXEL_FORMAT_RGBA_1010102: case HAL_PIXEL_FORMAT_ARGB_2101010: case HAL_PIXEL_FORMAT_BGRA_1010102: case HAL_PIXEL_FORMAT_ABGR_2101010: case HAL_PIXEL_FORMAT_RGBA_FP16: return true; default: return false; } } void GetRGBPlaneInfo(const BufferInfo &info, int32_t format, int32_t width, int32_t height, int32_t /* flags */, int *plane_count, PlaneLayoutInfo *plane_info) { uint64_t usage = info.usage; *plane_count = 1; uint32_t bpp = 0; if (IsUncompressedRGBFormat(format)) { bpp = GetBppForUncompressedRGB(format); } plane_info->component = (PlaneComponent)(PLANE_COMPONENT_R | PLANE_COMPONENT_G | PLANE_COMPONENT_B); if (HasAlphaComponent(format)) { plane_info->component = (PlaneComponent)(plane_info->component | PLANE_COMPONENT_A); } plane_info->size = GetSize(info, width, height); plane_info->step = bpp; plane_info->offset = GetRgbMetaSize(format, width, height, usage); plane_info->h_subsampling = 0; plane_info->v_subsampling = 0; plane_info->stride = width; plane_info->stride_bytes = width * plane_info->step; plane_info->scanlines = height; } // TODO(tbalacha): tile vs ubwc -- may need to find a diff way to differentiate void GetDRMFormat(uint32_t format, uint32_t flags, uint32_t *drm_format, uint64_t *drm_format_modifier) { bool compressed = (flags & private_handle_t::PRIV_FLAGS_UBWC_ALIGNED) ? true : false; switch (format) { case HAL_PIXEL_FORMAT_RGBA_8888: *drm_format = DRM_FORMAT_ABGR8888; break; case HAL_PIXEL_FORMAT_RGBA_5551: *drm_format = DRM_FORMAT_ABGR1555; break; case HAL_PIXEL_FORMAT_RGBA_4444: *drm_format = DRM_FORMAT_ABGR4444; break; case HAL_PIXEL_FORMAT_BGRA_8888: *drm_format = DRM_FORMAT_ARGB8888; break; case HAL_PIXEL_FORMAT_RGBX_8888: *drm_format = DRM_FORMAT_XBGR8888; if (compressed) *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; break; case HAL_PIXEL_FORMAT_BGRX_8888: *drm_format = DRM_FORMAT_XRGB8888; break; case HAL_PIXEL_FORMAT_RGB_888: *drm_format = DRM_FORMAT_BGR888; break; case HAL_PIXEL_FORMAT_RGB_565: *drm_format = DRM_FORMAT_BGR565; break; case HAL_PIXEL_FORMAT_BGR_565: *drm_format = DRM_FORMAT_BGR565; if (compressed) *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; break; case HAL_PIXEL_FORMAT_RGBA_1010102: *drm_format = DRM_FORMAT_ABGR2101010; if (compressed) *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; break; case HAL_PIXEL_FORMAT_ARGB_2101010: *drm_format = DRM_FORMAT_BGRA1010102; break; case HAL_PIXEL_FORMAT_RGBX_1010102: *drm_format = DRM_FORMAT_XBGR2101010; if (compressed) *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; break; case HAL_PIXEL_FORMAT_XRGB_2101010: *drm_format = DRM_FORMAT_BGRX1010102; break; case HAL_PIXEL_FORMAT_BGRA_1010102: *drm_format = DRM_FORMAT_ARGB2101010; break; case HAL_PIXEL_FORMAT_ABGR_2101010: *drm_format = DRM_FORMAT_RGBA1010102; break; case HAL_PIXEL_FORMAT_BGRX_1010102: *drm_format = DRM_FORMAT_XRGB2101010; break; case HAL_PIXEL_FORMAT_XBGR_2101010: *drm_format = DRM_FORMAT_RGBX1010102; break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: *drm_format = DRM_FORMAT_NV12; break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: *drm_format = DRM_FORMAT_NV12; if (compressed) { *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; } else { *drm_format_modifier = DRM_FORMAT_MOD_QCOM_TILE; } break; case HAL_PIXEL_FORMAT_YCrCb_420_SP: *drm_format = DRM_FORMAT_NV21; break; case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: *drm_format = DRM_FORMAT_NV21; break; case HAL_PIXEL_FORMAT_YCbCr_420_P010: case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: *drm_format = DRM_FORMAT_NV12; *drm_format_modifier = DRM_FORMAT_MOD_QCOM_DX; break; case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: *drm_format = DRM_FORMAT_NV12; if (compressed) { *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED | DRM_FORMAT_MOD_QCOM_DX; } else { *drm_format_modifier = DRM_FORMAT_MOD_QCOM_TILE | DRM_FORMAT_MOD_QCOM_DX; } break; case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: *drm_format = DRM_FORMAT_NV12; if (compressed) { *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED | DRM_FORMAT_MOD_QCOM_DX | DRM_FORMAT_MOD_QCOM_TIGHT; } else { *drm_format_modifier = DRM_FORMAT_MOD_QCOM_TILE | DRM_FORMAT_MOD_QCOM_DX | DRM_FORMAT_MOD_QCOM_TIGHT; } break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: *drm_format = DRM_FORMAT_NV16; break; /* TODO: No HAL_PIXEL_FORMAT equivalent? case kFormatYCrCb422H2V1SemiPlanar: *drm_format = DRM_FORMAT_NV61; break;*/ case HAL_PIXEL_FORMAT_YV12: *drm_format = DRM_FORMAT_YVU420; break; default: ALOGE("Unsupported format %d", format); } } } // namespace gralloc