You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

633 lines
24 KiB

/*
* Copyright (C) 2015 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//#define LOG_NDEBUG 0
#define LOG_TAG "ACameraMetadata"
#include "ACameraMetadata.h"
#include <utils/Vector.h>
#include <system/graphics.h>
#include <media/NdkImage.h>
using namespace android;
// Formats not listed in the public API, but still available to AImageReader
// Enum value must match corresponding enum in ui/PublicFormat.h (which is not
// available to VNDK)
enum AIMAGE_PRIVATE_FORMATS {
/**
* Unprocessed implementation-dependent raw
* depth measurements, opaque with 16 bit
* samples.
*
*/
AIMAGE_FORMAT_RAW_DEPTH = 0x1002,
/**
* Device specific 10 bits depth RAW image format.
*
* <p>Unprocessed implementation-dependent raw depth measurements, opaque with 10 bit samples
* and device specific bit layout.</p>
*/
AIMAGE_FORMAT_RAW_DEPTH10 = 0x1003,
};
/**
* ACameraMetadata Implementation
*/
ACameraMetadata::ACameraMetadata(camera_metadata_t* buffer, ACAMERA_METADATA_TYPE type) :
mData(std::make_shared<CameraMetadata>(buffer)),
mType(type) {
init();
}
ACameraMetadata::ACameraMetadata(const std::shared_ptr<CameraMetadata>& cameraMetadata,
ACAMERA_METADATA_TYPE type) :
mData(cameraMetadata),
mType(type) {
init();
}
ACameraMetadata::ACameraMetadata(const ACameraMetadata& other) :
mData(std::make_shared<CameraMetadata>(*(other.mData))),
mType(other.mType) {
}
ACameraMetadata::~ACameraMetadata() {
}
void
ACameraMetadata::init() {
if (mType == ACM_CHARACTERISTICS) {
filterUnsupportedFeatures();
filterStreamConfigurations();
filterDurations(ANDROID_SCALER_AVAILABLE_MIN_FRAME_DURATIONS);
filterDurations(ANDROID_SCALER_AVAILABLE_STALL_DURATIONS);
filterDurations(ANDROID_DEPTH_AVAILABLE_DEPTH_MIN_FRAME_DURATIONS);
filterDurations(ANDROID_DEPTH_AVAILABLE_DEPTH_STALL_DURATIONS);
filterDurations(ANDROID_HEIC_AVAILABLE_HEIC_MIN_FRAME_DURATIONS);
filterDurations(ANDROID_HEIC_AVAILABLE_HEIC_STALL_DURATIONS);
filterDurations(ANDROID_DEPTH_AVAILABLE_DYNAMIC_DEPTH_MIN_FRAME_DURATIONS);
filterDurations(ANDROID_DEPTH_AVAILABLE_DYNAMIC_DEPTH_STALL_DURATIONS);
}
// TODO: filter request/result keys
}
bool
ACameraMetadata::isNdkSupportedCapability(int32_t capability) {
switch (capability) {
case ANDROID_REQUEST_AVAILABLE_CAPABILITIES_YUV_REPROCESSING:
case ANDROID_REQUEST_AVAILABLE_CAPABILITIES_PRIVATE_REPROCESSING:
case ANDROID_REQUEST_AVAILABLE_CAPABILITIES_CONSTRAINED_HIGH_SPEED_VIDEO:
return false;
default:
// Assuming every capability passed to this function is actually a
// valid capability.
return true;
}
}
void
ACameraMetadata::filterUnsupportedFeatures() {
// Hide unsupported capabilities (reprocessing)
camera_metadata_entry entry = mData->find(ANDROID_REQUEST_AVAILABLE_CAPABILITIES);
if (entry.count == 0 || entry.type != TYPE_BYTE) {
ALOGE("%s: malformed available capability key! count %zu, type %d",
__FUNCTION__, entry.count, entry.type);
return;
}
Vector<uint8_t> capabilities;
capabilities.setCapacity(entry.count);
for (size_t i = 0; i < entry.count; i++) {
uint8_t capability = entry.data.u8[i];
if (isNdkSupportedCapability(capability)) {
capabilities.push(capability);
if (capability == ANDROID_REQUEST_AVAILABLE_CAPABILITIES_LOGICAL_MULTI_CAMERA) {
derivePhysicalCameraIds();
}
}
}
mData->update(ANDROID_REQUEST_AVAILABLE_CAPABILITIES, capabilities);
}
void
ACameraMetadata::derivePhysicalCameraIds() {
ACameraMetadata_const_entry entry;
auto ret = getConstEntry(ACAMERA_LOGICAL_MULTI_CAMERA_PHYSICAL_IDS, &entry);
if (ret != ACAMERA_OK) {
ALOGE("%s: Get ACAMERA_LOGICAL_MULTI_CAMERA_PHYSICAL_IDS key failed. ret %d",
__FUNCTION__, ret);
return;
}
const uint8_t* ids = entry.data.u8;
size_t start = 0;
for (size_t i = 0; i < entry.count; ++i) {
if (ids[i] == '\0') {
if (start != i) {
mStaticPhysicalCameraIdValues.push_back(String8((const char *)ids+start));
mStaticPhysicalCameraIds.push_back(mStaticPhysicalCameraIdValues.back().string());
}
start = i+1;
}
}
if (mStaticPhysicalCameraIds.size() < 2) {
ALOGW("%s: Logical multi-camera device only has %zu physical cameras",
__FUNCTION__, mStaticPhysicalCameraIds.size());
}
}
void
ACameraMetadata::filterDurations(uint32_t tag) {
const int STREAM_CONFIGURATION_SIZE = 4;
const int STREAM_FORMAT_OFFSET = 0;
const int STREAM_WIDTH_OFFSET = 1;
const int STREAM_HEIGHT_OFFSET = 2;
const int STREAM_DURATION_OFFSET = 3;
camera_metadata_entry entry = mData->find(tag);
if (entry.count == 0) {
// Duration keys can be missing when corresponding capture feature is not supported
return;
}
if (entry.count % 4 || entry.type != TYPE_INT64) {
ALOGE("%s: malformed duration key %d! count %zu, type %d",
__FUNCTION__, tag, entry.count, entry.type);
return;
}
Vector<int64_t> filteredDurations;
filteredDurations.setCapacity(entry.count * 2);
for (size_t i=0; i < entry.count; i += STREAM_CONFIGURATION_SIZE) {
int64_t format = entry.data.i64[i + STREAM_FORMAT_OFFSET];
int64_t width = entry.data.i64[i + STREAM_WIDTH_OFFSET];
int64_t height = entry.data.i64[i + STREAM_HEIGHT_OFFSET];
int64_t duration = entry.data.i32[i + STREAM_DURATION_OFFSET];
// Leave the unfiltered format in so apps depending on previous wrong
// filter behavior continue to work
filteredDurations.push_back(format);
filteredDurations.push_back(width);
filteredDurations.push_back(height);
filteredDurations.push_back(duration);
// Translate HAL formats to NDK format
switch (tag) {
case ANDROID_SCALER_AVAILABLE_MIN_FRAME_DURATIONS:
case ANDROID_SCALER_AVAILABLE_STALL_DURATIONS:
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_JPEG;
filteredDurations.push_back(format);
filteredDurations.push_back(width);
filteredDurations.push_back(height);
filteredDurations.push_back(duration);
}
break;
case ANDROID_DEPTH_AVAILABLE_DEPTH_MIN_FRAME_DURATIONS:
case ANDROID_DEPTH_AVAILABLE_DEPTH_STALL_DURATIONS:
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_DEPTH_POINT_CLOUD;
filteredDurations.push_back(format);
filteredDurations.push_back(width);
filteredDurations.push_back(height);
filteredDurations.push_back(duration);
} else if (format == HAL_PIXEL_FORMAT_Y16) {
format = AIMAGE_FORMAT_DEPTH16;
filteredDurations.push_back(format);
filteredDurations.push_back(width);
filteredDurations.push_back(height);
filteredDurations.push_back(duration);
}
break;
case ANDROID_HEIC_AVAILABLE_HEIC_MIN_FRAME_DURATIONS:
case ANDROID_HEIC_AVAILABLE_HEIC_STALL_DURATIONS:
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_HEIC;
filteredDurations.push_back(format);
filteredDurations.push_back(width);
filteredDurations.push_back(height);
filteredDurations.push_back(duration);
}
break;
case ANDROID_DEPTH_AVAILABLE_DYNAMIC_DEPTH_MIN_FRAME_DURATIONS:
case ANDROID_DEPTH_AVAILABLE_DYNAMIC_DEPTH_STALL_DURATIONS:
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_DEPTH_JPEG;
filteredDurations.push_back(format);
filteredDurations.push_back(width);
filteredDurations.push_back(height);
filteredDurations.push_back(duration);
}
break;
default:
// Should not reach here
ALOGE("%s: Unkown tag 0x%x", __FUNCTION__, tag);
}
}
mData->update(tag, filteredDurations);
}
void
ACameraMetadata::filterStreamConfigurations() {
const int STREAM_CONFIGURATION_SIZE = 4;
const int STREAM_FORMAT_OFFSET = 0;
const int STREAM_WIDTH_OFFSET = 1;
const int STREAM_HEIGHT_OFFSET = 2;
const int STREAM_IS_INPUT_OFFSET = 3;
camera_metadata_entry entry = mData->find(ANDROID_SCALER_AVAILABLE_STREAM_CONFIGURATIONS);
if (entry.count > 0 && (entry.count % 4 || entry.type != TYPE_INT32)) {
ALOGE("%s: malformed available stream configuration key! count %zu, type %d",
__FUNCTION__, entry.count, entry.type);
return;
}
Vector<int32_t> filteredStreamConfigs;
filteredStreamConfigs.setCapacity(entry.count);
for (size_t i=0; i < entry.count; i += STREAM_CONFIGURATION_SIZE) {
int32_t format = entry.data.i32[i + STREAM_FORMAT_OFFSET];
int32_t width = entry.data.i32[i + STREAM_WIDTH_OFFSET];
int32_t height = entry.data.i32[i + STREAM_HEIGHT_OFFSET];
int32_t isInput = entry.data.i32[i + STREAM_IS_INPUT_OFFSET];
if (isInput == ANDROID_SCALER_AVAILABLE_STREAM_CONFIGURATIONS_INPUT) {
// Hide input streams
continue;
}
// Translate HAL formats to NDK format
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_JPEG;
}
filteredStreamConfigs.push_back(format);
filteredStreamConfigs.push_back(width);
filteredStreamConfigs.push_back(height);
filteredStreamConfigs.push_back(isInput);
}
if (filteredStreamConfigs.size() > 0) {
mData->update(ANDROID_SCALER_AVAILABLE_STREAM_CONFIGURATIONS, filteredStreamConfigs);
}
entry = mData->find(ANDROID_DEPTH_AVAILABLE_DEPTH_STREAM_CONFIGURATIONS);
if (entry.count > 0 && (entry.count % 4 || entry.type != TYPE_INT32)) {
ALOGE("%s: malformed available depth stream configuration key! count %zu, type %d",
__FUNCTION__, entry.count, entry.type);
return;
}
Vector<int32_t> filteredDepthStreamConfigs;
filteredDepthStreamConfigs.setCapacity(entry.count);
for (size_t i=0; i < entry.count; i += STREAM_CONFIGURATION_SIZE) {
int32_t format = entry.data.i32[i + STREAM_FORMAT_OFFSET];
int32_t width = entry.data.i32[i + STREAM_WIDTH_OFFSET];
int32_t height = entry.data.i32[i + STREAM_HEIGHT_OFFSET];
int32_t isInput = entry.data.i32[i + STREAM_IS_INPUT_OFFSET];
if (isInput == ANDROID_SCALER_AVAILABLE_STREAM_CONFIGURATIONS_INPUT) {
// Hide input streams
continue;
}
// Translate HAL formats to NDK format
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_DEPTH_POINT_CLOUD;
} else if (format == HAL_PIXEL_FORMAT_Y16) {
format = AIMAGE_FORMAT_DEPTH16;
} else if (format == HAL_PIXEL_FORMAT_RAW16) {
format = static_cast<int32_t>(AIMAGE_FORMAT_RAW_DEPTH);
} else if (format == HAL_PIXEL_FORMAT_RAW10) {
format = static_cast<int32_t>(AIMAGE_FORMAT_RAW_DEPTH10);
}
filteredDepthStreamConfigs.push_back(format);
filteredDepthStreamConfigs.push_back(width);
filteredDepthStreamConfigs.push_back(height);
filteredDepthStreamConfigs.push_back(isInput);
}
if (filteredDepthStreamConfigs.size() > 0) {
mData->update(ANDROID_DEPTH_AVAILABLE_DEPTH_STREAM_CONFIGURATIONS,
filteredDepthStreamConfigs);
}
entry = mData->find(ANDROID_HEIC_AVAILABLE_HEIC_STREAM_CONFIGURATIONS);
Vector<int32_t> filteredHeicStreamConfigs;
filteredHeicStreamConfigs.setCapacity(entry.count);
for (size_t i=0; i < entry.count; i += STREAM_CONFIGURATION_SIZE) {
int32_t format = entry.data.i32[i + STREAM_FORMAT_OFFSET];
int32_t width = entry.data.i32[i + STREAM_WIDTH_OFFSET];
int32_t height = entry.data.i32[i + STREAM_HEIGHT_OFFSET];
int32_t isInput = entry.data.i32[i + STREAM_IS_INPUT_OFFSET];
if (isInput == ACAMERA_HEIC_AVAILABLE_HEIC_STREAM_CONFIGURATIONS_INPUT) {
// Hide input streams
continue;
}
// Translate HAL formats to NDK format
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_HEIC;
}
filteredHeicStreamConfigs.push_back(format);
filteredHeicStreamConfigs.push_back(width);
filteredHeicStreamConfigs.push_back(height);
filteredHeicStreamConfigs.push_back(isInput);
}
mData->update(ANDROID_HEIC_AVAILABLE_HEIC_STREAM_CONFIGURATIONS, filteredHeicStreamConfigs);
entry = mData->find(ANDROID_DEPTH_AVAILABLE_DYNAMIC_DEPTH_STREAM_CONFIGURATIONS);
Vector<int32_t> filteredDynamicDepthStreamConfigs;
filteredDynamicDepthStreamConfigs.setCapacity(entry.count);
for (size_t i = 0; i < entry.count; i += STREAM_CONFIGURATION_SIZE) {
int32_t format = entry.data.i32[i + STREAM_FORMAT_OFFSET];
int32_t width = entry.data.i32[i + STREAM_WIDTH_OFFSET];
int32_t height = entry.data.i32[i + STREAM_HEIGHT_OFFSET];
int32_t isInput = entry.data.i32[i + STREAM_IS_INPUT_OFFSET];
if (isInput == ACAMERA_DEPTH_AVAILABLE_DYNAMIC_DEPTH_STREAM_CONFIGURATIONS_INPUT) {
// Hide input streams
continue;
}
// Translate HAL formats to NDK format
if (format == HAL_PIXEL_FORMAT_BLOB) {
format = AIMAGE_FORMAT_DEPTH_JPEG;
}
filteredDynamicDepthStreamConfigs.push_back(format);
filteredDynamicDepthStreamConfigs.push_back(width);
filteredDynamicDepthStreamConfigs.push_back(height);
filteredDynamicDepthStreamConfigs.push_back(isInput);
}
mData->update(ACAMERA_DEPTH_AVAILABLE_DYNAMIC_DEPTH_STREAM_CONFIGURATIONS,
filteredDynamicDepthStreamConfigs);
}
bool
ACameraMetadata::isVendorTag(const uint32_t tag) {
uint32_t tag_section = tag >> 16;
if (tag_section >= VENDOR_SECTION) {
return true;
}
return false;
}
camera_status_t
ACameraMetadata::getConstEntry(uint32_t tag, ACameraMetadata_const_entry* entry) const {
if (entry == nullptr) {
return ACAMERA_ERROR_INVALID_PARAMETER;
}
Mutex::Autolock _l(mLock);
camera_metadata_ro_entry rawEntry = static_cast<const CameraMetadata*>(mData.get())->find(tag);
if (rawEntry.count == 0) {
ALOGE("%s: cannot find metadata tag %d", __FUNCTION__, tag);
return ACAMERA_ERROR_METADATA_NOT_FOUND;
}
entry->tag = tag;
entry->type = rawEntry.type;
entry->count = rawEntry.count;
entry->data.u8 = rawEntry.data.u8;
return ACAMERA_OK;
}
camera_status_t
ACameraMetadata::update(uint32_t tag, uint32_t count, const uint8_t* data) {
return updateImpl<uint8_t>(tag, count, data);
}
camera_status_t
ACameraMetadata::update(uint32_t tag, uint32_t count, const int32_t* data) {
return updateImpl<int32_t>(tag, count, data);
}
camera_status_t
ACameraMetadata::update(uint32_t tag, uint32_t count, const float* data) {
return updateImpl<float>(tag, count, data);
}
camera_status_t
ACameraMetadata::update(uint32_t tag, uint32_t count, const double* data) {
return updateImpl<double>(tag, count, data);
}
camera_status_t
ACameraMetadata::update(uint32_t tag, uint32_t count, const int64_t* data) {
return updateImpl<int64_t>(tag, count, data);
}
camera_status_t
ACameraMetadata::update(uint32_t tag, uint32_t count, const ACameraMetadata_rational* data) {
return updateImpl<camera_metadata_rational_t>(tag, count, data);
}
camera_status_t
ACameraMetadata::getTags(/*out*/int32_t* numTags,
/*out*/const uint32_t** tags) const {
Mutex::Autolock _l(mLock);
if (mTags.size() == 0) {
size_t entry_count = mData->entryCount();
mTags.setCapacity(entry_count);
const camera_metadata_t* rawMetadata = mData->getAndLock();
for (size_t i = 0; i < entry_count; i++) {
camera_metadata_ro_entry_t entry;
int ret = get_camera_metadata_ro_entry(rawMetadata, i, &entry);
if (ret != 0) {
mData->unlock(rawMetadata);
ALOGE("%s: error reading metadata index %zu", __FUNCTION__, i);
return ACAMERA_ERROR_UNKNOWN;
}
// Hide system key from users
if (sSystemTags.count(entry.tag) == 0) {
mTags.push_back(entry.tag);
}
}
mData->unlock(rawMetadata);
}
*numTags = mTags.size();
*tags = mTags.array();
return ACAMERA_OK;
}
const CameraMetadata&
ACameraMetadata::getInternalData() const {
return (*mData);
}
bool
ACameraMetadata::isLogicalMultiCamera(size_t* count, const char*const** physicalCameraIds) const {
if (mType != ACM_CHARACTERISTICS) {
ALOGE("%s must be called for a static metadata!", __FUNCTION__);
return false;
}
if (count == nullptr || physicalCameraIds == nullptr) {
ALOGE("%s: Invalid input count: %p, physicalCameraIds: %p", __FUNCTION__,
count, physicalCameraIds);
return false;
}
if (mStaticPhysicalCameraIds.size() >= 2) {
*count = mStaticPhysicalCameraIds.size();
*physicalCameraIds = mStaticPhysicalCameraIds.data();
return true;
}
return false;
}
// TODO: some of key below should be hidden from user
// ex: ACAMERA_REQUEST_ID and ACAMERA_REPROCESS_EFFECTIVE_EXPOSURE_FACTOR
/*@O~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~
* The key entries below this point are generated from metadata
* definitions in /system/media/camera/docs. Do not modify by hand or
* modify the comment blocks at the start or end.
*~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~*/
bool
ACameraMetadata::isCaptureRequestTag(const uint32_t tag) {
// Skip check for vendor keys
if (isVendorTag(tag)) {
return true;
}
switch (tag) {
case ACAMERA_COLOR_CORRECTION_MODE:
case ACAMERA_COLOR_CORRECTION_TRANSFORM:
case ACAMERA_COLOR_CORRECTION_GAINS:
case ACAMERA_COLOR_CORRECTION_ABERRATION_MODE:
case ACAMERA_CONTROL_AE_ANTIBANDING_MODE:
case ACAMERA_CONTROL_AE_EXPOSURE_COMPENSATION:
case ACAMERA_CONTROL_AE_LOCK:
case ACAMERA_CONTROL_AE_MODE:
case ACAMERA_CONTROL_AE_REGIONS:
case ACAMERA_CONTROL_AE_TARGET_FPS_RANGE:
case ACAMERA_CONTROL_AE_PRECAPTURE_TRIGGER:
case ACAMERA_CONTROL_AF_MODE:
case ACAMERA_CONTROL_AF_REGIONS:
case ACAMERA_CONTROL_AF_TRIGGER:
case ACAMERA_CONTROL_AWB_LOCK:
case ACAMERA_CONTROL_AWB_MODE:
case ACAMERA_CONTROL_AWB_REGIONS:
case ACAMERA_CONTROL_CAPTURE_INTENT:
case ACAMERA_CONTROL_EFFECT_MODE:
case ACAMERA_CONTROL_MODE:
case ACAMERA_CONTROL_SCENE_MODE:
case ACAMERA_CONTROL_VIDEO_STABILIZATION_MODE:
case ACAMERA_CONTROL_POST_RAW_SENSITIVITY_BOOST:
case ACAMERA_CONTROL_ENABLE_ZSL:
case ACAMERA_CONTROL_EXTENDED_SCENE_MODE:
case ACAMERA_CONTROL_ZOOM_RATIO:
case ACAMERA_EDGE_MODE:
case ACAMERA_FLASH_MODE:
case ACAMERA_HOT_PIXEL_MODE:
case ACAMERA_JPEG_GPS_COORDINATES:
case ACAMERA_JPEG_GPS_PROCESSING_METHOD:
case ACAMERA_JPEG_GPS_TIMESTAMP:
case ACAMERA_JPEG_ORIENTATION:
case ACAMERA_JPEG_QUALITY:
case ACAMERA_JPEG_THUMBNAIL_QUALITY:
case ACAMERA_JPEG_THUMBNAIL_SIZE:
case ACAMERA_LENS_APERTURE:
case ACAMERA_LENS_FILTER_DENSITY:
case ACAMERA_LENS_FOCAL_LENGTH:
case ACAMERA_LENS_FOCUS_DISTANCE:
case ACAMERA_LENS_OPTICAL_STABILIZATION_MODE:
case ACAMERA_NOISE_REDUCTION_MODE:
case ACAMERA_SCALER_CROP_REGION:
case ACAMERA_SCALER_ROTATE_AND_CROP:
case ACAMERA_SENSOR_EXPOSURE_TIME:
case ACAMERA_SENSOR_FRAME_DURATION:
case ACAMERA_SENSOR_SENSITIVITY:
case ACAMERA_SENSOR_TEST_PATTERN_DATA:
case ACAMERA_SENSOR_TEST_PATTERN_MODE:
case ACAMERA_SENSOR_PIXEL_MODE:
case ACAMERA_SHADING_MODE:
case ACAMERA_STATISTICS_FACE_DETECT_MODE:
case ACAMERA_STATISTICS_HOT_PIXEL_MAP_MODE:
case ACAMERA_STATISTICS_LENS_SHADING_MAP_MODE:
case ACAMERA_STATISTICS_OIS_DATA_MODE:
case ACAMERA_TONEMAP_CURVE_BLUE:
case ACAMERA_TONEMAP_CURVE_GREEN:
case ACAMERA_TONEMAP_CURVE_RED:
case ACAMERA_TONEMAP_MODE:
case ACAMERA_TONEMAP_GAMMA:
case ACAMERA_TONEMAP_PRESET_CURVE:
case ACAMERA_BLACK_LEVEL_LOCK:
case ACAMERA_DISTORTION_CORRECTION_MODE:
return true;
default:
return false;
}
}
// System tags that should be hidden from users
std::unordered_set<uint32_t> ACameraMetadata::sSystemTags ({
ANDROID_CONTROL_SCENE_MODE_OVERRIDES,
ANDROID_CONTROL_AE_PRECAPTURE_ID,
ANDROID_CONTROL_AF_TRIGGER_ID,
ANDROID_DEMOSAIC_MODE,
ANDROID_EDGE_STRENGTH,
ANDROID_FLASH_FIRING_POWER,
ANDROID_FLASH_FIRING_TIME,
ANDROID_FLASH_COLOR_TEMPERATURE,
ANDROID_FLASH_MAX_ENERGY,
ANDROID_FLASH_INFO_CHARGE_DURATION,
ANDROID_JPEG_MAX_SIZE,
ANDROID_JPEG_SIZE,
ANDROID_NOISE_REDUCTION_STRENGTH,
ANDROID_QUIRKS_METERING_CROP_REGION,
ANDROID_QUIRKS_TRIGGER_AF_WITH_AUTO,
ANDROID_QUIRKS_USE_ZSL_FORMAT,
ANDROID_REQUEST_INPUT_STREAMS,
ANDROID_REQUEST_METADATA_MODE,
ANDROID_REQUEST_OUTPUT_STREAMS,
ANDROID_REQUEST_TYPE,
ANDROID_REQUEST_MAX_NUM_REPROCESS_STREAMS,
ANDROID_SCALER_AVAILABLE_RAW_MIN_DURATIONS,
ANDROID_SCALER_AVAILABLE_RAW_SIZES,
ANDROID_SENSOR_BASE_GAIN_FACTOR,
ANDROID_SENSOR_PROFILE_HUE_SAT_MAP_DIMENSIONS,
ANDROID_SENSOR_TEMPERATURE,
ANDROID_SENSOR_PROFILE_HUE_SAT_MAP,
ANDROID_SENSOR_PROFILE_TONE_CURVE,
ANDROID_SENSOR_OPAQUE_RAW_SIZE,
ANDROID_SENSOR_OPAQUE_RAW_SIZE_MAXIMUM_RESOLUTION,
ANDROID_SHADING_STRENGTH,
ANDROID_STATISTICS_HISTOGRAM_MODE,
ANDROID_STATISTICS_SHARPNESS_MAP_MODE,
ANDROID_STATISTICS_HISTOGRAM,
ANDROID_STATISTICS_SHARPNESS_MAP,
ANDROID_STATISTICS_INFO_HISTOGRAM_BUCKET_COUNT,
ANDROID_STATISTICS_INFO_MAX_HISTOGRAM_COUNT,
ANDROID_STATISTICS_INFO_MAX_SHARPNESS_MAP_VALUE,
ANDROID_STATISTICS_INFO_SHARPNESS_MAP_SIZE,
ANDROID_INFO_SUPPORTED_BUFFER_MANAGEMENT_VERSION,
ANDROID_DEPTH_MAX_DEPTH_SAMPLES,
ANDROID_HEIC_INFO_SUPPORTED,
ANDROID_HEIC_INFO_MAX_JPEG_APP_SEGMENTS_COUNT,
});
/*~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~
* End generated code
*~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~O@*/