You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
1073 lines
43 KiB
1073 lines
43 KiB
/*
|
|
* Copyright 2017 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
//#define LOG_NDEBUG 0
|
|
#define LOG_TAG "C2SoftAvcDec"
|
|
#include <log/log.h>
|
|
|
|
#include <media/stagefright/foundation/MediaDefs.h>
|
|
|
|
#include <C2Debug.h>
|
|
#include <C2PlatformSupport.h>
|
|
#include <Codec2Mapper.h>
|
|
#include <SimpleC2Interface.h>
|
|
|
|
#include "C2SoftAvcDec.h"
|
|
|
|
namespace android {
|
|
|
|
namespace {
|
|
constexpr size_t kMinInputBufferSize = 2 * 1024 * 1024;
|
|
constexpr char COMPONENT_NAME[] = "c2.android.avc.decoder";
|
|
constexpr uint32_t kDefaultOutputDelay = 8;
|
|
/* avc specification allows for a maximum delay of 16 frames.
|
|
As soft avc decoder supports interlaced, this delay would be 32 fields.
|
|
And avc decoder implementation has an additional delay of 2 decode calls.
|
|
So total maximum output delay is 34 */
|
|
constexpr uint32_t kMaxOutputDelay = 34;
|
|
constexpr uint32_t kMinInputBytes = 4;
|
|
} // namespace
|
|
|
|
class C2SoftAvcDec::IntfImpl : public SimpleInterface<void>::BaseParams {
|
|
public:
|
|
explicit IntfImpl(const std::shared_ptr<C2ReflectorHelper> &helper)
|
|
: SimpleInterface<void>::BaseParams(
|
|
helper,
|
|
COMPONENT_NAME,
|
|
C2Component::KIND_DECODER,
|
|
C2Component::DOMAIN_VIDEO,
|
|
MEDIA_MIMETYPE_VIDEO_AVC) {
|
|
noPrivateBuffers(); // TODO: account for our buffers here
|
|
noInputReferences();
|
|
noOutputReferences();
|
|
noInputLatency();
|
|
noTimeStretch();
|
|
|
|
// TODO: Proper support for reorder depth.
|
|
addParameter(
|
|
DefineParam(mActualOutputDelay, C2_PARAMKEY_OUTPUT_DELAY)
|
|
.withDefault(new C2PortActualDelayTuning::output(kDefaultOutputDelay))
|
|
.withFields({C2F(mActualOutputDelay, value).inRange(0, kMaxOutputDelay)})
|
|
.withSetter(Setter<decltype(*mActualOutputDelay)>::StrictValueWithNoDeps)
|
|
.build());
|
|
|
|
// TODO: output latency and reordering
|
|
|
|
addParameter(
|
|
DefineParam(mAttrib, C2_PARAMKEY_COMPONENT_ATTRIBUTES)
|
|
.withConstValue(new C2ComponentAttributesSetting(C2Component::ATTRIB_IS_TEMPORAL))
|
|
.build());
|
|
|
|
// coded and output picture size is the same for this codec
|
|
addParameter(
|
|
DefineParam(mSize, C2_PARAMKEY_PICTURE_SIZE)
|
|
.withDefault(new C2StreamPictureSizeInfo::output(0u, 320, 240))
|
|
.withFields({
|
|
C2F(mSize, width).inRange(2, 4080, 2),
|
|
C2F(mSize, height).inRange(2, 4080, 2),
|
|
})
|
|
.withSetter(SizeSetter)
|
|
.build());
|
|
|
|
addParameter(
|
|
DefineParam(mMaxSize, C2_PARAMKEY_MAX_PICTURE_SIZE)
|
|
.withDefault(new C2StreamMaxPictureSizeTuning::output(0u, 320, 240))
|
|
.withFields({
|
|
C2F(mSize, width).inRange(2, 4080, 2),
|
|
C2F(mSize, height).inRange(2, 4080, 2),
|
|
})
|
|
.withSetter(MaxPictureSizeSetter, mSize)
|
|
.build());
|
|
|
|
addParameter(
|
|
DefineParam(mProfileLevel, C2_PARAMKEY_PROFILE_LEVEL)
|
|
.withDefault(new C2StreamProfileLevelInfo::input(0u,
|
|
C2Config::PROFILE_AVC_CONSTRAINED_BASELINE, C2Config::LEVEL_AVC_5_2))
|
|
.withFields({
|
|
C2F(mProfileLevel, profile).oneOf({
|
|
C2Config::PROFILE_AVC_CONSTRAINED_BASELINE,
|
|
C2Config::PROFILE_AVC_BASELINE,
|
|
C2Config::PROFILE_AVC_MAIN,
|
|
C2Config::PROFILE_AVC_CONSTRAINED_HIGH,
|
|
C2Config::PROFILE_AVC_PROGRESSIVE_HIGH,
|
|
C2Config::PROFILE_AVC_HIGH}),
|
|
C2F(mProfileLevel, level).oneOf({
|
|
C2Config::LEVEL_AVC_1, C2Config::LEVEL_AVC_1B, C2Config::LEVEL_AVC_1_1,
|
|
C2Config::LEVEL_AVC_1_2, C2Config::LEVEL_AVC_1_3,
|
|
C2Config::LEVEL_AVC_2, C2Config::LEVEL_AVC_2_1, C2Config::LEVEL_AVC_2_2,
|
|
C2Config::LEVEL_AVC_3, C2Config::LEVEL_AVC_3_1, C2Config::LEVEL_AVC_3_2,
|
|
C2Config::LEVEL_AVC_4, C2Config::LEVEL_AVC_4_1, C2Config::LEVEL_AVC_4_2,
|
|
C2Config::LEVEL_AVC_5, C2Config::LEVEL_AVC_5_1, C2Config::LEVEL_AVC_5_2
|
|
})
|
|
})
|
|
.withSetter(ProfileLevelSetter, mSize)
|
|
.build());
|
|
|
|
addParameter(
|
|
DefineParam(mMaxInputSize, C2_PARAMKEY_INPUT_MAX_BUFFER_SIZE)
|
|
.withDefault(new C2StreamMaxBufferSizeInfo::input(0u, kMinInputBufferSize))
|
|
.withFields({
|
|
C2F(mMaxInputSize, value).any(),
|
|
})
|
|
.calculatedAs(MaxInputSizeSetter, mMaxSize)
|
|
.build());
|
|
|
|
C2ChromaOffsetStruct locations[1] = { C2ChromaOffsetStruct::ITU_YUV_420_0() };
|
|
std::shared_ptr<C2StreamColorInfo::output> defaultColorInfo =
|
|
C2StreamColorInfo::output::AllocShared(
|
|
1u, 0u, 8u /* bitDepth */, C2Color::YUV_420);
|
|
memcpy(defaultColorInfo->m.locations, locations, sizeof(locations));
|
|
|
|
defaultColorInfo =
|
|
C2StreamColorInfo::output::AllocShared(
|
|
{ C2ChromaOffsetStruct::ITU_YUV_420_0() },
|
|
0u, 8u /* bitDepth */, C2Color::YUV_420);
|
|
helper->addStructDescriptors<C2ChromaOffsetStruct>();
|
|
|
|
addParameter(
|
|
DefineParam(mColorInfo, C2_PARAMKEY_CODED_COLOR_INFO)
|
|
.withConstValue(defaultColorInfo)
|
|
.build());
|
|
|
|
addParameter(
|
|
DefineParam(mDefaultColorAspects, C2_PARAMKEY_DEFAULT_COLOR_ASPECTS)
|
|
.withDefault(new C2StreamColorAspectsTuning::output(
|
|
0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED,
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
|
|
.withFields({
|
|
C2F(mDefaultColorAspects, range).inRange(
|
|
C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
|
|
C2F(mDefaultColorAspects, primaries).inRange(
|
|
C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER),
|
|
C2F(mDefaultColorAspects, transfer).inRange(
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::TRANSFER_OTHER),
|
|
C2F(mDefaultColorAspects, matrix).inRange(
|
|
C2Color::MATRIX_UNSPECIFIED, C2Color::MATRIX_OTHER)
|
|
})
|
|
.withSetter(DefaultColorAspectsSetter)
|
|
.build());
|
|
|
|
addParameter(
|
|
DefineParam(mCodedColorAspects, C2_PARAMKEY_VUI_COLOR_ASPECTS)
|
|
.withDefault(new C2StreamColorAspectsInfo::input(
|
|
0u, C2Color::RANGE_LIMITED, C2Color::PRIMARIES_UNSPECIFIED,
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
|
|
.withFields({
|
|
C2F(mCodedColorAspects, range).inRange(
|
|
C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
|
|
C2F(mCodedColorAspects, primaries).inRange(
|
|
C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER),
|
|
C2F(mCodedColorAspects, transfer).inRange(
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::TRANSFER_OTHER),
|
|
C2F(mCodedColorAspects, matrix).inRange(
|
|
C2Color::MATRIX_UNSPECIFIED, C2Color::MATRIX_OTHER)
|
|
})
|
|
.withSetter(CodedColorAspectsSetter)
|
|
.build());
|
|
|
|
addParameter(
|
|
DefineParam(mColorAspects, C2_PARAMKEY_COLOR_ASPECTS)
|
|
.withDefault(new C2StreamColorAspectsInfo::output(
|
|
0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED,
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED))
|
|
.withFields({
|
|
C2F(mColorAspects, range).inRange(
|
|
C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER),
|
|
C2F(mColorAspects, primaries).inRange(
|
|
C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER),
|
|
C2F(mColorAspects, transfer).inRange(
|
|
C2Color::TRANSFER_UNSPECIFIED, C2Color::TRANSFER_OTHER),
|
|
C2F(mColorAspects, matrix).inRange(
|
|
C2Color::MATRIX_UNSPECIFIED, C2Color::MATRIX_OTHER)
|
|
})
|
|
.withSetter(ColorAspectsSetter, mDefaultColorAspects, mCodedColorAspects)
|
|
.build());
|
|
|
|
// TODO: support more formats?
|
|
addParameter(
|
|
DefineParam(mPixelFormat, C2_PARAMKEY_PIXEL_FORMAT)
|
|
.withConstValue(new C2StreamPixelFormatInfo::output(
|
|
0u, HAL_PIXEL_FORMAT_YCBCR_420_888))
|
|
.build());
|
|
}
|
|
static C2R SizeSetter(bool mayBlock, const C2P<C2StreamPictureSizeInfo::output> &oldMe,
|
|
C2P<C2StreamPictureSizeInfo::output> &me) {
|
|
(void)mayBlock;
|
|
C2R res = C2R::Ok();
|
|
if (!me.F(me.v.width).supportsAtAll(me.v.width)) {
|
|
res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.width)));
|
|
me.set().width = oldMe.v.width;
|
|
}
|
|
if (!me.F(me.v.height).supportsAtAll(me.v.height)) {
|
|
res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.height)));
|
|
me.set().height = oldMe.v.height;
|
|
}
|
|
return res;
|
|
}
|
|
|
|
static C2R MaxPictureSizeSetter(bool mayBlock, C2P<C2StreamMaxPictureSizeTuning::output> &me,
|
|
const C2P<C2StreamPictureSizeInfo::output> &size) {
|
|
(void)mayBlock;
|
|
// TODO: get max width/height from the size's field helpers vs. hardcoding
|
|
me.set().width = c2_min(c2_max(me.v.width, size.v.width), 4080u);
|
|
me.set().height = c2_min(c2_max(me.v.height, size.v.height), 4080u);
|
|
return C2R::Ok();
|
|
}
|
|
|
|
static C2R MaxInputSizeSetter(bool mayBlock, C2P<C2StreamMaxBufferSizeInfo::input> &me,
|
|
const C2P<C2StreamMaxPictureSizeTuning::output> &maxSize) {
|
|
(void)mayBlock;
|
|
// assume compression ratio of 2
|
|
me.set().value = c2_max((((maxSize.v.width + 15) / 16)
|
|
* ((maxSize.v.height + 15) / 16) * 192), kMinInputBufferSize);
|
|
return C2R::Ok();
|
|
}
|
|
|
|
static C2R ProfileLevelSetter(bool mayBlock, C2P<C2StreamProfileLevelInfo::input> &me,
|
|
const C2P<C2StreamPictureSizeInfo::output> &size) {
|
|
(void)mayBlock;
|
|
(void)size;
|
|
(void)me; // TODO: validate
|
|
return C2R::Ok();
|
|
}
|
|
|
|
static C2R DefaultColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsTuning::output> &me) {
|
|
(void)mayBlock;
|
|
if (me.v.range > C2Color::RANGE_OTHER) {
|
|
me.set().range = C2Color::RANGE_OTHER;
|
|
}
|
|
if (me.v.primaries > C2Color::PRIMARIES_OTHER) {
|
|
me.set().primaries = C2Color::PRIMARIES_OTHER;
|
|
}
|
|
if (me.v.transfer > C2Color::TRANSFER_OTHER) {
|
|
me.set().transfer = C2Color::TRANSFER_OTHER;
|
|
}
|
|
if (me.v.matrix > C2Color::MATRIX_OTHER) {
|
|
me.set().matrix = C2Color::MATRIX_OTHER;
|
|
}
|
|
return C2R::Ok();
|
|
}
|
|
|
|
static C2R CodedColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsInfo::input> &me) {
|
|
(void)mayBlock;
|
|
if (me.v.range > C2Color::RANGE_OTHER) {
|
|
me.set().range = C2Color::RANGE_OTHER;
|
|
}
|
|
if (me.v.primaries > C2Color::PRIMARIES_OTHER) {
|
|
me.set().primaries = C2Color::PRIMARIES_OTHER;
|
|
}
|
|
if (me.v.transfer > C2Color::TRANSFER_OTHER) {
|
|
me.set().transfer = C2Color::TRANSFER_OTHER;
|
|
}
|
|
if (me.v.matrix > C2Color::MATRIX_OTHER) {
|
|
me.set().matrix = C2Color::MATRIX_OTHER;
|
|
}
|
|
return C2R::Ok();
|
|
}
|
|
|
|
static C2R ColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsInfo::output> &me,
|
|
const C2P<C2StreamColorAspectsTuning::output> &def,
|
|
const C2P<C2StreamColorAspectsInfo::input> &coded) {
|
|
(void)mayBlock;
|
|
// take default values for all unspecified fields, and coded values for specified ones
|
|
me.set().range = coded.v.range == RANGE_UNSPECIFIED ? def.v.range : coded.v.range;
|
|
me.set().primaries = coded.v.primaries == PRIMARIES_UNSPECIFIED
|
|
? def.v.primaries : coded.v.primaries;
|
|
me.set().transfer = coded.v.transfer == TRANSFER_UNSPECIFIED
|
|
? def.v.transfer : coded.v.transfer;
|
|
me.set().matrix = coded.v.matrix == MATRIX_UNSPECIFIED ? def.v.matrix : coded.v.matrix;
|
|
return C2R::Ok();
|
|
}
|
|
|
|
std::shared_ptr<C2StreamColorAspectsInfo::output> getColorAspects_l() {
|
|
return mColorAspects;
|
|
}
|
|
|
|
private:
|
|
std::shared_ptr<C2StreamProfileLevelInfo::input> mProfileLevel;
|
|
std::shared_ptr<C2StreamPictureSizeInfo::output> mSize;
|
|
std::shared_ptr<C2StreamMaxPictureSizeTuning::output> mMaxSize;
|
|
std::shared_ptr<C2StreamMaxBufferSizeInfo::input> mMaxInputSize;
|
|
std::shared_ptr<C2StreamColorInfo::output> mColorInfo;
|
|
std::shared_ptr<C2StreamColorAspectsInfo::input> mCodedColorAspects;
|
|
std::shared_ptr<C2StreamColorAspectsTuning::output> mDefaultColorAspects;
|
|
std::shared_ptr<C2StreamColorAspectsInfo::output> mColorAspects;
|
|
std::shared_ptr<C2StreamPixelFormatInfo::output> mPixelFormat;
|
|
};
|
|
|
|
static size_t getCpuCoreCount() {
|
|
long cpuCoreCount = 1;
|
|
#if defined(_SC_NPROCESSORS_ONLN)
|
|
cpuCoreCount = sysconf(_SC_NPROCESSORS_ONLN);
|
|
#else
|
|
// _SC_NPROC_ONLN must be defined...
|
|
cpuCoreCount = sysconf(_SC_NPROC_ONLN);
|
|
#endif
|
|
CHECK(cpuCoreCount >= 1);
|
|
ALOGV("Number of CPU cores: %ld", cpuCoreCount);
|
|
return (size_t)cpuCoreCount;
|
|
}
|
|
|
|
static void *ivd_aligned_malloc(void *ctxt, WORD32 alignment, WORD32 size) {
|
|
(void) ctxt;
|
|
return memalign(alignment, size);
|
|
}
|
|
|
|
static void ivd_aligned_free(void *ctxt, void *mem) {
|
|
(void) ctxt;
|
|
free(mem);
|
|
}
|
|
|
|
C2SoftAvcDec::C2SoftAvcDec(
|
|
const char *name,
|
|
c2_node_id_t id,
|
|
const std::shared_ptr<IntfImpl> &intfImpl)
|
|
: SimpleC2Component(std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)),
|
|
mIntf(intfImpl),
|
|
mDecHandle(nullptr),
|
|
mOutBufferFlush(nullptr),
|
|
mIvColorFormat(IV_YUV_420P),
|
|
mOutputDelay(kDefaultOutputDelay),
|
|
mWidth(320),
|
|
mHeight(240),
|
|
mHeaderDecoded(false),
|
|
mOutIndex(0u) {
|
|
GENERATE_FILE_NAMES();
|
|
CREATE_DUMP_FILE(mInFile);
|
|
}
|
|
|
|
C2SoftAvcDec::~C2SoftAvcDec() {
|
|
onRelease();
|
|
}
|
|
|
|
c2_status_t C2SoftAvcDec::onInit() {
|
|
status_t err = initDecoder();
|
|
return err == OK ? C2_OK : C2_CORRUPTED;
|
|
}
|
|
|
|
c2_status_t C2SoftAvcDec::onStop() {
|
|
if (OK != resetDecoder()) return C2_CORRUPTED;
|
|
resetPlugin();
|
|
return C2_OK;
|
|
}
|
|
|
|
void C2SoftAvcDec::onReset() {
|
|
(void) onStop();
|
|
}
|
|
|
|
void C2SoftAvcDec::onRelease() {
|
|
(void) deleteDecoder();
|
|
if (mOutBufferFlush) {
|
|
ivd_aligned_free(nullptr, mOutBufferFlush);
|
|
mOutBufferFlush = nullptr;
|
|
}
|
|
if (mOutBlock) {
|
|
mOutBlock.reset();
|
|
}
|
|
}
|
|
|
|
c2_status_t C2SoftAvcDec::onFlush_sm() {
|
|
if (OK != setFlushMode()) return C2_CORRUPTED;
|
|
|
|
uint32_t bufferSize = mStride * mHeight * 3 / 2;
|
|
mOutBufferFlush = (uint8_t *)ivd_aligned_malloc(nullptr, 128, bufferSize);
|
|
if (!mOutBufferFlush) {
|
|
ALOGE("could not allocate tmp output buffer (for flush) of size %u ", bufferSize);
|
|
return C2_NO_MEMORY;
|
|
}
|
|
|
|
while (true) {
|
|
ih264d_video_decode_ip_t s_h264d_decode_ip = {};
|
|
ih264d_video_decode_op_t s_h264d_decode_op = {};
|
|
ivd_video_decode_ip_t *ps_decode_ip = &s_h264d_decode_ip.s_ivd_video_decode_ip_t;
|
|
ivd_video_decode_op_t *ps_decode_op = &s_h264d_decode_op.s_ivd_video_decode_op_t;
|
|
|
|
setDecodeArgs(ps_decode_ip, ps_decode_op, nullptr, nullptr, 0, 0, 0);
|
|
(void) ivdec_api_function(mDecHandle, &s_h264d_decode_ip, &s_h264d_decode_op);
|
|
if (0 == ps_decode_op->u4_output_present) {
|
|
resetPlugin();
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (mOutBufferFlush) {
|
|
ivd_aligned_free(nullptr, mOutBufferFlush);
|
|
mOutBufferFlush = nullptr;
|
|
}
|
|
|
|
return C2_OK;
|
|
}
|
|
|
|
status_t C2SoftAvcDec::createDecoder() {
|
|
ivdext_create_ip_t s_create_ip = {};
|
|
ivdext_create_op_t s_create_op = {};
|
|
|
|
s_create_ip.s_ivd_create_ip_t.u4_size = sizeof(ivdext_create_ip_t);
|
|
s_create_ip.s_ivd_create_ip_t.e_cmd = IVD_CMD_CREATE;
|
|
s_create_ip.s_ivd_create_ip_t.u4_share_disp_buf = 0;
|
|
s_create_ip.s_ivd_create_ip_t.e_output_format = mIvColorFormat;
|
|
s_create_ip.s_ivd_create_ip_t.pf_aligned_alloc = ivd_aligned_malloc;
|
|
s_create_ip.s_ivd_create_ip_t.pf_aligned_free = ivd_aligned_free;
|
|
s_create_ip.s_ivd_create_ip_t.pv_mem_ctxt = nullptr;
|
|
s_create_op.s_ivd_create_op_t.u4_size = sizeof(ivdext_create_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(nullptr,
|
|
&s_create_ip,
|
|
&s_create_op);
|
|
if (status != IV_SUCCESS) {
|
|
ALOGE("error in %s: 0x%x", __func__,
|
|
s_create_op.s_ivd_create_op_t.u4_error_code);
|
|
return UNKNOWN_ERROR;
|
|
}
|
|
mDecHandle = (iv_obj_t*)s_create_op.s_ivd_create_op_t.pv_handle;
|
|
mDecHandle->pv_fxns = (void *)ivdec_api_function;
|
|
mDecHandle->u4_size = sizeof(iv_obj_t);
|
|
|
|
return OK;
|
|
}
|
|
|
|
status_t C2SoftAvcDec::setNumCores() {
|
|
ivdext_ctl_set_num_cores_ip_t s_set_num_cores_ip = {};
|
|
ivdext_ctl_set_num_cores_op_t s_set_num_cores_op = {};
|
|
|
|
s_set_num_cores_ip.u4_size = sizeof(ivdext_ctl_set_num_cores_ip_t);
|
|
s_set_num_cores_ip.e_cmd = IVD_CMD_VIDEO_CTL;
|
|
s_set_num_cores_ip.e_sub_cmd = IVDEXT_CMD_CTL_SET_NUM_CORES;
|
|
s_set_num_cores_ip.u4_num_cores = mNumCores;
|
|
s_set_num_cores_op.u4_size = sizeof(ivdext_ctl_set_num_cores_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_set_num_cores_ip,
|
|
&s_set_num_cores_op);
|
|
if (IV_SUCCESS != status) {
|
|
ALOGD("error in %s: 0x%x", __func__, s_set_num_cores_op.u4_error_code);
|
|
return UNKNOWN_ERROR;
|
|
}
|
|
|
|
return OK;
|
|
}
|
|
|
|
status_t C2SoftAvcDec::setParams(size_t stride, IVD_VIDEO_DECODE_MODE_T dec_mode) {
|
|
ih264d_ctl_set_config_ip_t s_h264d_set_dyn_params_ip = {};
|
|
ih264d_ctl_set_config_op_t s_h264d_set_dyn_params_op = {};
|
|
ivd_ctl_set_config_ip_t *ps_set_dyn_params_ip =
|
|
&s_h264d_set_dyn_params_ip.s_ivd_ctl_set_config_ip_t;
|
|
ivd_ctl_set_config_op_t *ps_set_dyn_params_op =
|
|
&s_h264d_set_dyn_params_op.s_ivd_ctl_set_config_op_t;
|
|
|
|
ps_set_dyn_params_ip->u4_size = sizeof(ih264d_ctl_set_config_ip_t);
|
|
ps_set_dyn_params_ip->e_cmd = IVD_CMD_VIDEO_CTL;
|
|
ps_set_dyn_params_ip->e_sub_cmd = IVD_CMD_CTL_SETPARAMS;
|
|
ps_set_dyn_params_ip->u4_disp_wd = (UWORD32) stride;
|
|
ps_set_dyn_params_ip->e_frm_skip_mode = IVD_SKIP_NONE;
|
|
ps_set_dyn_params_ip->e_frm_out_mode = IVD_DISPLAY_FRAME_OUT;
|
|
ps_set_dyn_params_ip->e_vid_dec_mode = dec_mode;
|
|
ps_set_dyn_params_op->u4_size = sizeof(ih264d_ctl_set_config_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_h264d_set_dyn_params_ip,
|
|
&s_h264d_set_dyn_params_op);
|
|
if (status != IV_SUCCESS) {
|
|
ALOGE("error in %s: 0x%x", __func__, ps_set_dyn_params_op->u4_error_code);
|
|
return UNKNOWN_ERROR;
|
|
}
|
|
|
|
return OK;
|
|
}
|
|
|
|
void C2SoftAvcDec::getVersion() {
|
|
ivd_ctl_getversioninfo_ip_t s_get_versioninfo_ip = {};
|
|
ivd_ctl_getversioninfo_op_t s_get_versioninfo_op = {};
|
|
UWORD8 au1_buf[512];
|
|
|
|
s_get_versioninfo_ip.u4_size = sizeof(ivd_ctl_getversioninfo_ip_t);
|
|
s_get_versioninfo_ip.e_cmd = IVD_CMD_VIDEO_CTL;
|
|
s_get_versioninfo_ip.e_sub_cmd = IVD_CMD_CTL_GETVERSION;
|
|
s_get_versioninfo_ip.pv_version_buffer = au1_buf;
|
|
s_get_versioninfo_ip.u4_version_buffer_size = sizeof(au1_buf);
|
|
s_get_versioninfo_op.u4_size = sizeof(ivd_ctl_getversioninfo_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_get_versioninfo_ip,
|
|
&s_get_versioninfo_op);
|
|
if (status != IV_SUCCESS) {
|
|
ALOGD("error in %s: 0x%x", __func__,
|
|
s_get_versioninfo_op.u4_error_code);
|
|
} else {
|
|
ALOGV("ittiam decoder version number: %s",
|
|
(char *) s_get_versioninfo_ip.pv_version_buffer);
|
|
}
|
|
}
|
|
|
|
status_t C2SoftAvcDec::initDecoder() {
|
|
if (OK != createDecoder()) return UNKNOWN_ERROR;
|
|
mNumCores = MIN(getCpuCoreCount(), MAX_NUM_CORES);
|
|
mStride = ALIGN32(mWidth);
|
|
mSignalledError = false;
|
|
resetPlugin();
|
|
(void) setNumCores();
|
|
if (OK != setParams(mStride, IVD_DECODE_FRAME)) return UNKNOWN_ERROR;
|
|
(void) getVersion();
|
|
|
|
return OK;
|
|
}
|
|
|
|
bool C2SoftAvcDec::setDecodeArgs(ivd_video_decode_ip_t *ps_decode_ip,
|
|
ivd_video_decode_op_t *ps_decode_op,
|
|
C2ReadView *inBuffer,
|
|
C2GraphicView *outBuffer,
|
|
size_t inOffset,
|
|
size_t inSize,
|
|
uint32_t tsMarker) {
|
|
uint32_t displayStride = mStride;
|
|
if (outBuffer) {
|
|
C2PlanarLayout layout;
|
|
layout = outBuffer->layout();
|
|
displayStride = layout.planes[C2PlanarLayout::PLANE_Y].rowInc;
|
|
}
|
|
uint32_t displayHeight = mHeight;
|
|
size_t lumaSize = displayStride * displayHeight;
|
|
size_t chromaSize = lumaSize >> 2;
|
|
|
|
if (mStride != displayStride) {
|
|
mStride = displayStride;
|
|
if (OK != setParams(mStride, IVD_DECODE_FRAME)) return false;
|
|
}
|
|
|
|
ps_decode_ip->u4_size = sizeof(ih264d_video_decode_ip_t);
|
|
ps_decode_ip->e_cmd = IVD_CMD_VIDEO_DECODE;
|
|
if (inBuffer) {
|
|
ps_decode_ip->u4_ts = tsMarker;
|
|
ps_decode_ip->pv_stream_buffer = const_cast<uint8_t *>(inBuffer->data() + inOffset);
|
|
ps_decode_ip->u4_num_Bytes = inSize;
|
|
} else {
|
|
ps_decode_ip->u4_ts = 0;
|
|
ps_decode_ip->pv_stream_buffer = nullptr;
|
|
ps_decode_ip->u4_num_Bytes = 0;
|
|
}
|
|
ps_decode_ip->s_out_buffer.u4_min_out_buf_size[0] = lumaSize;
|
|
ps_decode_ip->s_out_buffer.u4_min_out_buf_size[1] = chromaSize;
|
|
ps_decode_ip->s_out_buffer.u4_min_out_buf_size[2] = chromaSize;
|
|
if (outBuffer) {
|
|
if (outBuffer->height() < displayHeight) {
|
|
ALOGE("Output buffer too small: provided (%dx%d) required (%ux%u)",
|
|
outBuffer->width(), outBuffer->height(), displayStride, displayHeight);
|
|
return false;
|
|
}
|
|
ps_decode_ip->s_out_buffer.pu1_bufs[0] = outBuffer->data()[C2PlanarLayout::PLANE_Y];
|
|
ps_decode_ip->s_out_buffer.pu1_bufs[1] = outBuffer->data()[C2PlanarLayout::PLANE_U];
|
|
ps_decode_ip->s_out_buffer.pu1_bufs[2] = outBuffer->data()[C2PlanarLayout::PLANE_V];
|
|
} else {
|
|
ps_decode_ip->s_out_buffer.pu1_bufs[0] = mOutBufferFlush;
|
|
ps_decode_ip->s_out_buffer.pu1_bufs[1] = mOutBufferFlush + lumaSize;
|
|
ps_decode_ip->s_out_buffer.pu1_bufs[2] = mOutBufferFlush + lumaSize + chromaSize;
|
|
}
|
|
ps_decode_ip->s_out_buffer.u4_num_bufs = 3;
|
|
ps_decode_op->u4_size = sizeof(ih264d_video_decode_op_t);
|
|
|
|
return true;
|
|
}
|
|
|
|
bool C2SoftAvcDec::getVuiParams() {
|
|
ivdext_ctl_get_vui_params_ip_t s_get_vui_params_ip = {};
|
|
ivdext_ctl_get_vui_params_op_t s_get_vui_params_op = {};
|
|
|
|
s_get_vui_params_ip.u4_size = sizeof(ivdext_ctl_get_vui_params_ip_t);
|
|
s_get_vui_params_ip.e_cmd = IVD_CMD_VIDEO_CTL;
|
|
s_get_vui_params_ip.e_sub_cmd =
|
|
(IVD_CONTROL_API_COMMAND_TYPE_T) IH264D_CMD_CTL_GET_VUI_PARAMS;
|
|
s_get_vui_params_op.u4_size = sizeof(ivdext_ctl_get_vui_params_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_get_vui_params_ip,
|
|
&s_get_vui_params_op);
|
|
if (status != IV_SUCCESS) {
|
|
ALOGD("error in %s: 0x%x", __func__, s_get_vui_params_op.u4_error_code);
|
|
return false;
|
|
}
|
|
|
|
VuiColorAspects vuiColorAspects;
|
|
vuiColorAspects.primaries = s_get_vui_params_op.u1_colour_primaries;
|
|
vuiColorAspects.transfer = s_get_vui_params_op.u1_tfr_chars;
|
|
vuiColorAspects.coeffs = s_get_vui_params_op.u1_matrix_coeffs;
|
|
vuiColorAspects.fullRange = s_get_vui_params_op.u1_video_full_range_flag;
|
|
|
|
// convert vui aspects to C2 values if changed
|
|
if (!(vuiColorAspects == mBitstreamColorAspects)) {
|
|
mBitstreamColorAspects = vuiColorAspects;
|
|
ColorAspects sfAspects;
|
|
C2StreamColorAspectsInfo::input codedAspects = { 0u };
|
|
ColorUtils::convertIsoColorAspectsToCodecAspects(
|
|
vuiColorAspects.primaries, vuiColorAspects.transfer, vuiColorAspects.coeffs,
|
|
vuiColorAspects.fullRange, sfAspects);
|
|
if (!C2Mapper::map(sfAspects.mPrimaries, &codedAspects.primaries)) {
|
|
codedAspects.primaries = C2Color::PRIMARIES_UNSPECIFIED;
|
|
}
|
|
if (!C2Mapper::map(sfAspects.mRange, &codedAspects.range)) {
|
|
codedAspects.range = C2Color::RANGE_UNSPECIFIED;
|
|
}
|
|
if (!C2Mapper::map(sfAspects.mMatrixCoeffs, &codedAspects.matrix)) {
|
|
codedAspects.matrix = C2Color::MATRIX_UNSPECIFIED;
|
|
}
|
|
if (!C2Mapper::map(sfAspects.mTransfer, &codedAspects.transfer)) {
|
|
codedAspects.transfer = C2Color::TRANSFER_UNSPECIFIED;
|
|
}
|
|
std::vector<std::unique_ptr<C2SettingResult>> failures;
|
|
(void)mIntf->config({&codedAspects}, C2_MAY_BLOCK, &failures);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
status_t C2SoftAvcDec::setFlushMode() {
|
|
ivd_ctl_flush_ip_t s_set_flush_ip = {};
|
|
ivd_ctl_flush_op_t s_set_flush_op = {};
|
|
|
|
s_set_flush_ip.u4_size = sizeof(ivd_ctl_flush_ip_t);
|
|
s_set_flush_ip.e_cmd = IVD_CMD_VIDEO_CTL;
|
|
s_set_flush_ip.e_sub_cmd = IVD_CMD_CTL_FLUSH;
|
|
s_set_flush_op.u4_size = sizeof(ivd_ctl_flush_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_set_flush_ip,
|
|
&s_set_flush_op);
|
|
if (status != IV_SUCCESS) {
|
|
ALOGE("error in %s: 0x%x", __func__, s_set_flush_op.u4_error_code);
|
|
return UNKNOWN_ERROR;
|
|
}
|
|
|
|
return OK;
|
|
}
|
|
|
|
status_t C2SoftAvcDec::resetDecoder() {
|
|
ivd_ctl_reset_ip_t s_reset_ip = {};
|
|
ivd_ctl_reset_op_t s_reset_op = {};
|
|
|
|
s_reset_ip.u4_size = sizeof(ivd_ctl_reset_ip_t);
|
|
s_reset_ip.e_cmd = IVD_CMD_VIDEO_CTL;
|
|
s_reset_ip.e_sub_cmd = IVD_CMD_CTL_RESET;
|
|
s_reset_op.u4_size = sizeof(ivd_ctl_reset_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_reset_ip,
|
|
&s_reset_op);
|
|
if (IV_SUCCESS != status) {
|
|
ALOGE("error in %s: 0x%x", __func__, s_reset_op.u4_error_code);
|
|
return UNKNOWN_ERROR;
|
|
}
|
|
mStride = 0;
|
|
(void) setNumCores();
|
|
mSignalledError = false;
|
|
mHeaderDecoded = false;
|
|
|
|
return OK;
|
|
}
|
|
|
|
void C2SoftAvcDec::resetPlugin() {
|
|
mSignalledOutputEos = false;
|
|
gettimeofday(&mTimeStart, nullptr);
|
|
gettimeofday(&mTimeEnd, nullptr);
|
|
}
|
|
|
|
status_t C2SoftAvcDec::deleteDecoder() {
|
|
if (mDecHandle) {
|
|
ivdext_delete_ip_t s_delete_ip = {};
|
|
ivdext_delete_op_t s_delete_op = {};
|
|
|
|
s_delete_ip.s_ivd_delete_ip_t.u4_size = sizeof(ivdext_delete_ip_t);
|
|
s_delete_ip.s_ivd_delete_ip_t.e_cmd = IVD_CMD_DELETE;
|
|
s_delete_op.s_ivd_delete_op_t.u4_size = sizeof(ivdext_delete_op_t);
|
|
IV_API_CALL_STATUS_T status = ivdec_api_function(mDecHandle,
|
|
&s_delete_ip,
|
|
&s_delete_op);
|
|
if (status != IV_SUCCESS) {
|
|
ALOGE("error in %s: 0x%x", __func__,
|
|
s_delete_op.s_ivd_delete_op_t.u4_error_code);
|
|
return UNKNOWN_ERROR;
|
|
}
|
|
mDecHandle = nullptr;
|
|
}
|
|
|
|
return OK;
|
|
}
|
|
|
|
static void fillEmptyWork(const std::unique_ptr<C2Work> &work) {
|
|
uint32_t flags = 0;
|
|
if (work->input.flags & C2FrameData::FLAG_END_OF_STREAM) {
|
|
flags |= C2FrameData::FLAG_END_OF_STREAM;
|
|
ALOGV("signalling eos");
|
|
}
|
|
work->worklets.front()->output.flags = (C2FrameData::flags_t)flags;
|
|
work->worklets.front()->output.buffers.clear();
|
|
work->worklets.front()->output.ordinal = work->input.ordinal;
|
|
work->workletsProcessed = 1u;
|
|
}
|
|
|
|
void C2SoftAvcDec::finishWork(uint64_t index, const std::unique_ptr<C2Work> &work) {
|
|
std::shared_ptr<C2Buffer> buffer = createGraphicBuffer(std::move(mOutBlock),
|
|
C2Rect(mWidth, mHeight));
|
|
mOutBlock = nullptr;
|
|
{
|
|
IntfImpl::Lock lock = mIntf->lock();
|
|
buffer->setInfo(mIntf->getColorAspects_l());
|
|
}
|
|
|
|
class FillWork {
|
|
public:
|
|
FillWork(uint32_t flags, C2WorkOrdinalStruct ordinal,
|
|
const std::shared_ptr<C2Buffer>& buffer)
|
|
: mFlags(flags), mOrdinal(ordinal), mBuffer(buffer) {}
|
|
~FillWork() = default;
|
|
|
|
void operator()(const std::unique_ptr<C2Work>& work) {
|
|
work->worklets.front()->output.flags = (C2FrameData::flags_t)mFlags;
|
|
work->worklets.front()->output.buffers.clear();
|
|
work->worklets.front()->output.ordinal = mOrdinal;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_OK;
|
|
if (mBuffer) {
|
|
work->worklets.front()->output.buffers.push_back(mBuffer);
|
|
}
|
|
ALOGV("timestamp = %lld, index = %lld, w/%s buffer",
|
|
mOrdinal.timestamp.peekll(), mOrdinal.frameIndex.peekll(),
|
|
mBuffer ? "" : "o");
|
|
}
|
|
|
|
private:
|
|
const uint32_t mFlags;
|
|
const C2WorkOrdinalStruct mOrdinal;
|
|
const std::shared_ptr<C2Buffer> mBuffer;
|
|
};
|
|
|
|
auto fillWork = [buffer](const std::unique_ptr<C2Work> &work) {
|
|
work->worklets.front()->output.flags = (C2FrameData::flags_t)0;
|
|
work->worklets.front()->output.buffers.clear();
|
|
work->worklets.front()->output.buffers.push_back(buffer);
|
|
work->worklets.front()->output.ordinal = work->input.ordinal;
|
|
work->workletsProcessed = 1u;
|
|
};
|
|
if (work && c2_cntr64_t(index) == work->input.ordinal.frameIndex) {
|
|
bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0);
|
|
// TODO: Check if cloneAndSend can be avoided by tracking number of frames remaining
|
|
if (eos) {
|
|
if (buffer) {
|
|
mOutIndex = index;
|
|
C2WorkOrdinalStruct outOrdinal = work->input.ordinal;
|
|
cloneAndSend(
|
|
mOutIndex, work,
|
|
FillWork(C2FrameData::FLAG_INCOMPLETE, outOrdinal, buffer));
|
|
buffer.reset();
|
|
}
|
|
} else {
|
|
fillWork(work);
|
|
}
|
|
} else {
|
|
finish(index, fillWork);
|
|
}
|
|
}
|
|
|
|
c2_status_t C2SoftAvcDec::ensureDecoderState(const std::shared_ptr<C2BlockPool> &pool) {
|
|
if (!mDecHandle) {
|
|
ALOGE("not supposed to be here, invalid decoder context");
|
|
return C2_CORRUPTED;
|
|
}
|
|
if (mOutBlock &&
|
|
(mOutBlock->width() != ALIGN32(mWidth) || mOutBlock->height() != mHeight)) {
|
|
mOutBlock.reset();
|
|
}
|
|
if (!mOutBlock) {
|
|
uint32_t format = HAL_PIXEL_FORMAT_YV12;
|
|
C2MemoryUsage usage = { C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE };
|
|
c2_status_t err =
|
|
pool->fetchGraphicBlock(ALIGN32(mWidth), mHeight, format, usage, &mOutBlock);
|
|
if (err != C2_OK) {
|
|
ALOGE("fetchGraphicBlock for Output failed with status %d", err);
|
|
return err;
|
|
}
|
|
ALOGV("provided (%dx%d) required (%dx%d)",
|
|
mOutBlock->width(), mOutBlock->height(), ALIGN32(mWidth), mHeight);
|
|
}
|
|
|
|
return C2_OK;
|
|
}
|
|
|
|
// TODO: can overall error checking be improved?
|
|
// TODO: allow configuration of color format and usage for graphic buffers instead
|
|
// of hard coding them to HAL_PIXEL_FORMAT_YV12
|
|
// TODO: pass coloraspects information to surface
|
|
// TODO: test support for dynamic change in resolution
|
|
// TODO: verify if the decoder sent back all frames
|
|
void C2SoftAvcDec::process(
|
|
const std::unique_ptr<C2Work> &work,
|
|
const std::shared_ptr<C2BlockPool> &pool) {
|
|
// Initialize output work
|
|
work->result = C2_OK;
|
|
work->workletsProcessed = 0u;
|
|
work->worklets.front()->output.flags = work->input.flags;
|
|
if (mSignalledError || mSignalledOutputEos) {
|
|
work->result = C2_BAD_VALUE;
|
|
return;
|
|
}
|
|
|
|
size_t inOffset = 0u;
|
|
size_t inSize = 0u;
|
|
uint32_t workIndex = work->input.ordinal.frameIndex.peeku() & 0xFFFFFFFF;
|
|
C2ReadView rView = mDummyReadView;
|
|
if (!work->input.buffers.empty()) {
|
|
rView = work->input.buffers[0]->data().linearBlocks().front().map().get();
|
|
inSize = rView.capacity();
|
|
if (inSize && rView.error()) {
|
|
ALOGE("read view map failed %d", rView.error());
|
|
work->result = rView.error();
|
|
return;
|
|
}
|
|
}
|
|
bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0);
|
|
bool hasPicture = false;
|
|
|
|
ALOGV("in buffer attr. size %zu timestamp %d frameindex %d, flags %x",
|
|
inSize, (int)work->input.ordinal.timestamp.peeku(),
|
|
(int)work->input.ordinal.frameIndex.peeku(), work->input.flags);
|
|
size_t inPos = 0;
|
|
while (inPos < inSize && inSize - inPos >= kMinInputBytes) {
|
|
if (C2_OK != ensureDecoderState(pool)) {
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
}
|
|
|
|
ih264d_video_decode_ip_t s_h264d_decode_ip = {};
|
|
ih264d_video_decode_op_t s_h264d_decode_op = {};
|
|
ivd_video_decode_ip_t *ps_decode_ip = &s_h264d_decode_ip.s_ivd_video_decode_ip_t;
|
|
ivd_video_decode_op_t *ps_decode_op = &s_h264d_decode_op.s_ivd_video_decode_op_t;
|
|
{
|
|
C2GraphicView wView = mOutBlock->map().get();
|
|
if (wView.error()) {
|
|
ALOGE("graphic view map failed %d", wView.error());
|
|
work->result = wView.error();
|
|
return;
|
|
}
|
|
if (!setDecodeArgs(ps_decode_ip, ps_decode_op, &rView, &wView,
|
|
inOffset + inPos, inSize - inPos, workIndex)) {
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
}
|
|
|
|
if (false == mHeaderDecoded) {
|
|
/* Decode header and get dimensions */
|
|
setParams(mStride, IVD_DECODE_HEADER);
|
|
}
|
|
|
|
WORD32 delay;
|
|
GETTIME(&mTimeStart, nullptr);
|
|
TIME_DIFF(mTimeEnd, mTimeStart, delay);
|
|
(void) ivdec_api_function(mDecHandle, &s_h264d_decode_ip, &s_h264d_decode_op);
|
|
WORD32 decodeTime;
|
|
GETTIME(&mTimeEnd, nullptr);
|
|
TIME_DIFF(mTimeStart, mTimeEnd, decodeTime);
|
|
ALOGV("decodeTime=%6d delay=%6d numBytes=%6d", decodeTime, delay,
|
|
ps_decode_op->u4_num_bytes_consumed);
|
|
}
|
|
if (IVD_MEM_ALLOC_FAILED == (ps_decode_op->u4_error_code & IVD_ERROR_MASK)) {
|
|
ALOGE("allocation failure in decoder");
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
} else if (IVD_STREAM_WIDTH_HEIGHT_NOT_SUPPORTED ==
|
|
(ps_decode_op->u4_error_code & IVD_ERROR_MASK)) {
|
|
ALOGE("unsupported resolution : %dx%d", mWidth, mHeight);
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
} else if (IVD_RES_CHANGED == (ps_decode_op->u4_error_code & IVD_ERROR_MASK)) {
|
|
ALOGV("resolution changed");
|
|
drainInternal(DRAIN_COMPONENT_NO_EOS, pool, work);
|
|
resetDecoder();
|
|
resetPlugin();
|
|
work->workletsProcessed = 0u;
|
|
|
|
/* Decode header and get new dimensions */
|
|
setParams(mStride, IVD_DECODE_HEADER);
|
|
(void) ivdec_api_function(mDecHandle, ps_decode_ip, ps_decode_op);
|
|
} else if (IS_IVD_FATAL_ERROR(ps_decode_op->u4_error_code)) {
|
|
ALOGE("Fatal error in decoder 0x%x", ps_decode_op->u4_error_code);
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
}
|
|
if (ps_decode_op->i4_reorder_depth >= 0 && mOutputDelay != ps_decode_op->i4_reorder_depth) {
|
|
mOutputDelay = ps_decode_op->i4_reorder_depth;
|
|
ALOGV("New Output delay %d ", mOutputDelay);
|
|
|
|
C2PortActualDelayTuning::output outputDelay(mOutputDelay);
|
|
std::vector<std::unique_ptr<C2SettingResult>> failures;
|
|
c2_status_t err =
|
|
mIntf->config({&outputDelay}, C2_MAY_BLOCK, &failures);
|
|
if (err == OK) {
|
|
work->worklets.front()->output.configUpdate.push_back(
|
|
C2Param::Copy(outputDelay));
|
|
} else {
|
|
ALOGE("Cannot set output delay");
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
}
|
|
}
|
|
if (0 < ps_decode_op->u4_pic_wd && 0 < ps_decode_op->u4_pic_ht) {
|
|
if (mHeaderDecoded == false) {
|
|
mHeaderDecoded = true;
|
|
mStride = ALIGN32(ps_decode_op->u4_pic_wd);
|
|
setParams(mStride, IVD_DECODE_FRAME);
|
|
}
|
|
if (ps_decode_op->u4_pic_wd != mWidth || ps_decode_op->u4_pic_ht != mHeight) {
|
|
mWidth = ps_decode_op->u4_pic_wd;
|
|
mHeight = ps_decode_op->u4_pic_ht;
|
|
CHECK_EQ(0u, ps_decode_op->u4_output_present);
|
|
|
|
C2StreamPictureSizeInfo::output size(0u, mWidth, mHeight);
|
|
std::vector<std::unique_ptr<C2SettingResult>> failures;
|
|
c2_status_t err = mIntf->config({&size}, C2_MAY_BLOCK, &failures);
|
|
if (err == OK) {
|
|
work->worklets.front()->output.configUpdate.push_back(
|
|
C2Param::Copy(size));
|
|
} else {
|
|
ALOGE("Cannot set width and height");
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return;
|
|
}
|
|
continue;
|
|
}
|
|
}
|
|
(void)getVuiParams();
|
|
hasPicture |= (1 == ps_decode_op->u4_frame_decoded_flag);
|
|
if (ps_decode_op->u4_output_present) {
|
|
finishWork(ps_decode_op->u4_ts, work);
|
|
}
|
|
inPos += ps_decode_op->u4_num_bytes_consumed;
|
|
}
|
|
if (eos) {
|
|
drainInternal(DRAIN_COMPONENT_WITH_EOS, pool, work);
|
|
mSignalledOutputEos = true;
|
|
} else if (!hasPicture) {
|
|
fillEmptyWork(work);
|
|
}
|
|
|
|
work->input.buffers.clear();
|
|
}
|
|
|
|
c2_status_t C2SoftAvcDec::drainInternal(
|
|
uint32_t drainMode,
|
|
const std::shared_ptr<C2BlockPool> &pool,
|
|
const std::unique_ptr<C2Work> &work) {
|
|
if (drainMode == NO_DRAIN) {
|
|
ALOGW("drain with NO_DRAIN: no-op");
|
|
return C2_OK;
|
|
}
|
|
if (drainMode == DRAIN_CHAIN) {
|
|
ALOGW("DRAIN_CHAIN not supported");
|
|
return C2_OMITTED;
|
|
}
|
|
|
|
if (OK != setFlushMode()) return C2_CORRUPTED;
|
|
while (true) {
|
|
if (C2_OK != ensureDecoderState(pool)) {
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
work->result = C2_CORRUPTED;
|
|
return C2_CORRUPTED;
|
|
}
|
|
C2GraphicView wView = mOutBlock->map().get();
|
|
if (wView.error()) {
|
|
ALOGE("graphic view map failed %d", wView.error());
|
|
return C2_CORRUPTED;
|
|
}
|
|
ih264d_video_decode_ip_t s_h264d_decode_ip = {};
|
|
ih264d_video_decode_op_t s_h264d_decode_op = {};
|
|
ivd_video_decode_ip_t *ps_decode_ip = &s_h264d_decode_ip.s_ivd_video_decode_ip_t;
|
|
ivd_video_decode_op_t *ps_decode_op = &s_h264d_decode_op.s_ivd_video_decode_op_t;
|
|
if (!setDecodeArgs(ps_decode_ip, ps_decode_op, nullptr, &wView, 0, 0, 0)) {
|
|
mSignalledError = true;
|
|
work->workletsProcessed = 1u;
|
|
return C2_CORRUPTED;
|
|
}
|
|
(void) ivdec_api_function(mDecHandle, &s_h264d_decode_ip, &s_h264d_decode_op);
|
|
if (ps_decode_op->u4_output_present) {
|
|
finishWork(ps_decode_op->u4_ts, work);
|
|
} else {
|
|
fillEmptyWork(work);
|
|
break;
|
|
}
|
|
}
|
|
|
|
return C2_OK;
|
|
}
|
|
|
|
c2_status_t C2SoftAvcDec::drain(
|
|
uint32_t drainMode,
|
|
const std::shared_ptr<C2BlockPool> &pool) {
|
|
return drainInternal(drainMode, pool, nullptr);
|
|
}
|
|
|
|
class C2SoftAvcDecFactory : public C2ComponentFactory {
|
|
public:
|
|
C2SoftAvcDecFactory() : mHelper(std::static_pointer_cast<C2ReflectorHelper>(
|
|
GetCodec2PlatformComponentStore()->getParamReflector())) {
|
|
}
|
|
|
|
virtual c2_status_t createComponent(
|
|
c2_node_id_t id,
|
|
std::shared_ptr<C2Component>* const component,
|
|
std::function<void(C2Component*)> deleter) override {
|
|
*component = std::shared_ptr<C2Component>(
|
|
new C2SoftAvcDec(COMPONENT_NAME,
|
|
id,
|
|
std::make_shared<C2SoftAvcDec::IntfImpl>(mHelper)),
|
|
deleter);
|
|
return C2_OK;
|
|
}
|
|
|
|
virtual c2_status_t createInterface(
|
|
c2_node_id_t id,
|
|
std::shared_ptr<C2ComponentInterface>* const interface,
|
|
std::function<void(C2ComponentInterface*)> deleter) override {
|
|
*interface = std::shared_ptr<C2ComponentInterface>(
|
|
new SimpleInterface<C2SoftAvcDec::IntfImpl>(
|
|
COMPONENT_NAME, id, std::make_shared<C2SoftAvcDec::IntfImpl>(mHelper)),
|
|
deleter);
|
|
return C2_OK;
|
|
}
|
|
|
|
virtual ~C2SoftAvcDecFactory() override = default;
|
|
|
|
private:
|
|
std::shared_ptr<C2ReflectorHelper> mHelper;
|
|
};
|
|
|
|
} // namespace android
|
|
|
|
__attribute__((cfi_canonical_jump_table))
|
|
extern "C" ::C2ComponentFactory* CreateCodec2Factory() {
|
|
ALOGV("in %s", __func__);
|
|
return new ::android::C2SoftAvcDecFactory();
|
|
}
|
|
|
|
__attribute__((cfi_canonical_jump_table))
|
|
extern "C" void DestroyCodec2Factory(::C2ComponentFactory* factory) {
|
|
ALOGV("in %s", __func__);
|
|
delete factory;
|
|
}
|