You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

653 lines
27 KiB

/*
* Copyright (C) 2020 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
// #define LOG_NDEBUG 0
#define LOG_TAG "VideoTrackTranscoder"
#include <android-base/logging.h>
#include <android-base/properties.h>
#include <media/NdkCommon.h>
#include <media/VideoTrackTranscoder.h>
#include <sys/prctl.h>
using namespace AMediaFormatUtils;
namespace android {
// Check that the codec sample flags have the expected NDK meaning.
static_assert(SAMPLE_FLAG_CODEC_CONFIG == AMEDIACODEC_BUFFER_FLAG_CODEC_CONFIG,
"Sample flag mismatch: CODEC_CONFIG");
static_assert(SAMPLE_FLAG_END_OF_STREAM == AMEDIACODEC_BUFFER_FLAG_END_OF_STREAM,
"Sample flag mismatch: END_OF_STREAM");
static_assert(SAMPLE_FLAG_PARTIAL_FRAME == AMEDIACODEC_BUFFER_FLAG_PARTIAL_FRAME,
"Sample flag mismatch: PARTIAL_FRAME");
// Color format defined by surface. (See MediaCodecInfo.CodecCapabilities#COLOR_FormatSurface.)
static constexpr int32_t kColorFormatSurface = 0x7f000789;
// Default key frame interval in seconds.
static constexpr float kDefaultKeyFrameIntervalSeconds = 1.0f;
// Default codec operating rate.
static int32_t kDefaultCodecOperatingRate720P = base::GetIntProperty(
"debug.media.transcoding.codec_max_operating_rate_720P", /*default*/ 480);
static int32_t kDefaultCodecOperatingRate1080P = base::GetIntProperty(
"debug.media.transcoding.codec_max_operating_rate_1080P", /*default*/ 240);
// Default codec priority.
static constexpr int32_t kDefaultCodecPriority = 1;
// Default bitrate, in case source estimation fails.
static constexpr int32_t kDefaultBitrateMbps = 10 * 1000 * 1000;
// Default frame rate.
static constexpr int32_t kDefaultFrameRate = 30;
// Default codec complexity
static constexpr int32_t kDefaultCodecComplexity = 1;
template <typename T>
void VideoTrackTranscoder::BlockingQueue<T>::push(T const& value, bool front) {
{
std::scoped_lock lock(mMutex);
if (mAborted) {
return;
}
if (front) {
mQueue.push_front(value);
} else {
mQueue.push_back(value);
}
}
mCondition.notify_one();
}
template <typename T>
T VideoTrackTranscoder::BlockingQueue<T>::pop() {
std::unique_lock lock(mMutex);
while (mQueue.empty()) {
mCondition.wait(lock);
}
T value = mQueue.front();
mQueue.pop_front();
return value;
}
// Note: Do not call if another thread might waiting in pop.
template <typename T>
void VideoTrackTranscoder::BlockingQueue<T>::abort() {
std::scoped_lock lock(mMutex);
mAborted = true;
mQueue.clear();
}
// The CodecWrapper class is used to let AMediaCodec instances outlive the transcoder object itself
// by giving the codec a weak pointer to the transcoder. Codecs wrapped in this object are kept
// alive by the transcoder and the codec's outstanding buffers. Once the transcoder stops and all
// output buffers have been released by downstream components the codec will also be released.
class VideoTrackTranscoder::CodecWrapper {
public:
CodecWrapper(AMediaCodec* codec, const std::weak_ptr<VideoTrackTranscoder>& transcoder)
: mCodec(codec), mTranscoder(transcoder), mCodecStarted(false) {}
~CodecWrapper() {
if (mCodecStarted) {
AMediaCodec_stop(mCodec);
}
AMediaCodec_delete(mCodec);
}
AMediaCodec* getCodec() { return mCodec; }
std::shared_ptr<VideoTrackTranscoder> getTranscoder() const { return mTranscoder.lock(); };
void setStarted() { mCodecStarted = true; }
private:
AMediaCodec* mCodec;
std::weak_ptr<VideoTrackTranscoder> mTranscoder;
bool mCodecStarted;
};
// Dispatch responses to codec callbacks onto the message queue.
struct AsyncCodecCallbackDispatch {
static void onAsyncInputAvailable(AMediaCodec* codec, void* userdata, int32_t index) {
VideoTrackTranscoder::CodecWrapper* wrapper =
static_cast<VideoTrackTranscoder::CodecWrapper*>(userdata);
if (auto transcoder = wrapper->getTranscoder()) {
if (codec == transcoder->mDecoder) {
transcoder->mCodecMessageQueue.push(
[transcoder, index] { transcoder->enqueueInputSample(index); });
}
}
}
static void onAsyncOutputAvailable(AMediaCodec* codec, void* userdata, int32_t index,
AMediaCodecBufferInfo* bufferInfoPtr) {
VideoTrackTranscoder::CodecWrapper* wrapper =
static_cast<VideoTrackTranscoder::CodecWrapper*>(userdata);
AMediaCodecBufferInfo bufferInfo = *bufferInfoPtr;
if (auto transcoder = wrapper->getTranscoder()) {
transcoder->mCodecMessageQueue.push([transcoder, index, codec, bufferInfo] {
if (codec == transcoder->mDecoder) {
transcoder->transferBuffer(index, bufferInfo);
} else if (codec == transcoder->mEncoder->getCodec()) {
transcoder->dequeueOutputSample(index, bufferInfo);
}
});
}
}
static void onAsyncFormatChanged(AMediaCodec* codec, void* userdata, AMediaFormat* format) {
VideoTrackTranscoder::CodecWrapper* wrapper =
static_cast<VideoTrackTranscoder::CodecWrapper*>(userdata);
if (auto transcoder = wrapper->getTranscoder()) {
const bool isDecoder = codec == transcoder->mDecoder;
const char* kCodecName = (isDecoder ? "Decoder" : "Encoder");
LOG(INFO) << kCodecName << " format changed: " << AMediaFormat_toString(format);
transcoder->mCodecMessageQueue.push([transcoder, format, isDecoder] {
transcoder->updateTrackFormat(format, isDecoder);
});
}
}
static void onAsyncError(AMediaCodec* codec, void* userdata, media_status_t error,
int32_t actionCode, const char* detail) {
LOG(ERROR) << "Error from codec " << codec << ", userdata " << userdata << ", error "
<< error << ", action " << actionCode << ", detail " << detail;
VideoTrackTranscoder::CodecWrapper* wrapper =
static_cast<VideoTrackTranscoder::CodecWrapper*>(userdata);
if (auto transcoder = wrapper->getTranscoder()) {
transcoder->mCodecMessageQueue.push(
[transcoder, error] { transcoder->mStatus = error; }, true);
}
}
};
// static
std::shared_ptr<VideoTrackTranscoder> VideoTrackTranscoder::create(
const std::weak_ptr<MediaTrackTranscoderCallback>& transcoderCallback, pid_t pid,
uid_t uid) {
return std::shared_ptr<VideoTrackTranscoder>(
new VideoTrackTranscoder(transcoderCallback, pid, uid));
}
VideoTrackTranscoder::~VideoTrackTranscoder() {
if (mDecoder != nullptr) {
AMediaCodec_delete(mDecoder);
}
if (mSurface != nullptr) {
ANativeWindow_release(mSurface);
}
}
// Search the default operating rate based on resolution.
static int32_t getDefaultOperatingRate(AMediaFormat* encoderFormat) {
int32_t width, height;
if (AMediaFormat_getInt32(encoderFormat, AMEDIAFORMAT_KEY_WIDTH, &width) && (width > 0) &&
AMediaFormat_getInt32(encoderFormat, AMEDIAFORMAT_KEY_HEIGHT, &height) && (height > 0)) {
if ((width == 1280 && height == 720) || (width == 720 && height == 1280)) {
return kDefaultCodecOperatingRate720P;
} else if ((width == 1920 && height == 1080) || (width == 1080 && height == 1920)) {
return kDefaultCodecOperatingRate1080P;
} else {
LOG(WARNING) << "Could not find default operating rate: " << width << " " << height;
// Don't set operating rate if the correct dimensions are not found.
}
} else {
LOG(ERROR) << "Failed to get default operating rate due to missing resolution";
}
return -1;
}
// Creates and configures the codecs.
media_status_t VideoTrackTranscoder::configureDestinationFormat(
const std::shared_ptr<AMediaFormat>& destinationFormat) {
media_status_t status = AMEDIA_OK;
if (destinationFormat == nullptr) {
LOG(ERROR) << "Destination format is null, use passthrough transcoder";
return AMEDIA_ERROR_INVALID_PARAMETER;
}
AMediaFormat* encoderFormat = AMediaFormat_new();
if (!encoderFormat || AMediaFormat_copy(encoderFormat, destinationFormat.get()) != AMEDIA_OK) {
LOG(ERROR) << "Unable to copy destination format";
return AMEDIA_ERROR_INVALID_PARAMETER;
}
if (!AMediaFormat_getInt32(encoderFormat, AMEDIAFORMAT_KEY_BIT_RATE, &mConfiguredBitrate)) {
status = mMediaSampleReader->getEstimatedBitrateForTrack(mTrackIndex, &mConfiguredBitrate);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to estimate bitrate. Using default " << kDefaultBitrateMbps;
mConfiguredBitrate = kDefaultBitrateMbps;
}
LOG(INFO) << "Configuring bitrate " << mConfiguredBitrate;
AMediaFormat_setInt32(encoderFormat, AMEDIAFORMAT_KEY_BIT_RATE, mConfiguredBitrate);
}
SetDefaultFormatValueFloat(AMEDIAFORMAT_KEY_I_FRAME_INTERVAL, encoderFormat,
kDefaultKeyFrameIntervalSeconds);
int32_t operatingRate = getDefaultOperatingRate(encoderFormat);
if (operatingRate != -1) {
float tmpf;
int32_t tmpi;
if (!AMediaFormat_getFloat(encoderFormat, AMEDIAFORMAT_KEY_OPERATING_RATE, &tmpf) &&
!AMediaFormat_getInt32(encoderFormat, AMEDIAFORMAT_KEY_OPERATING_RATE, &tmpi)) {
AMediaFormat_setInt32(encoderFormat, AMEDIAFORMAT_KEY_OPERATING_RATE, operatingRate);
}
}
SetDefaultFormatValueInt32(AMEDIAFORMAT_KEY_PRIORITY, encoderFormat, kDefaultCodecPriority);
SetDefaultFormatValueInt32(AMEDIAFORMAT_KEY_FRAME_RATE, encoderFormat, kDefaultFrameRate);
SetDefaultFormatValueInt32(AMEDIAFORMAT_KEY_COMPLEXITY, encoderFormat, kDefaultCodecComplexity);
AMediaFormat_setInt32(encoderFormat, AMEDIAFORMAT_KEY_COLOR_FORMAT, kColorFormatSurface);
// Always encode without rotation. The rotation degree will be transferred directly to
// MediaSampleWriter track format, and MediaSampleWriter will call AMediaMuxer_setOrientationHint.
AMediaFormat_setInt32(encoderFormat, AMEDIAFORMAT_KEY_ROTATION, 0);
// Request encoder to use background priorities by default.
SetDefaultFormatValueInt32(TBD_AMEDIACODEC_PARAMETER_KEY_BACKGROUND_MODE, encoderFormat,
1 /* true */);
mDestinationFormat = std::shared_ptr<AMediaFormat>(encoderFormat, &AMediaFormat_delete);
// Create and configure the encoder.
const char* destinationMime = nullptr;
bool ok = AMediaFormat_getString(mDestinationFormat.get(), AMEDIAFORMAT_KEY_MIME,
&destinationMime);
if (!ok) {
LOG(ERROR) << "Destination MIME type is required for transcoding.";
return AMEDIA_ERROR_INVALID_PARAMETER;
}
#define __TRANSCODING_MIN_API__ 31
AMediaCodec* encoder;
if (__builtin_available(android __TRANSCODING_MIN_API__, *)) {
encoder = AMediaCodec_createEncoderByTypeForClient(destinationMime, mPid, mUid);
} else {
encoder = AMediaCodec_createEncoderByType(destinationMime);
}
if (encoder == nullptr) {
LOG(ERROR) << "Unable to create encoder for type " << destinationMime;
return AMEDIA_ERROR_UNSUPPORTED;
}
mEncoder = std::make_shared<CodecWrapper>(encoder, shared_from_this());
LOG(INFO) << "Configuring encoder with: " << AMediaFormat_toString(mDestinationFormat.get());
status = AMediaCodec_configure(mEncoder->getCodec(), mDestinationFormat.get(),
NULL /* surface */, NULL /* crypto */,
AMEDIACODEC_CONFIGURE_FLAG_ENCODE);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to configure video encoder: " << status;
return status;
}
status = AMediaCodec_createInputSurface(mEncoder->getCodec(), &mSurface);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to create an encoder input surface: %d" << status;
return status;
}
// Create and configure the decoder.
const char* sourceMime = nullptr;
ok = AMediaFormat_getString(mSourceFormat.get(), AMEDIAFORMAT_KEY_MIME, &sourceMime);
if (!ok) {
LOG(ERROR) << "Source MIME type is required for transcoding.";
return AMEDIA_ERROR_INVALID_PARAMETER;
}
if (__builtin_available(android __TRANSCODING_MIN_API__, *)) {
mDecoder = AMediaCodec_createDecoderByTypeForClient(sourceMime, mPid, mUid);
} else {
mDecoder = AMediaCodec_createDecoderByType(sourceMime);
}
if (mDecoder == nullptr) {
LOG(ERROR) << "Unable to create decoder for type " << sourceMime;
return AMEDIA_ERROR_UNSUPPORTED;
}
auto decoderFormat = std::shared_ptr<AMediaFormat>(AMediaFormat_new(), &AMediaFormat_delete);
if (!decoderFormat ||
AMediaFormat_copy(decoderFormat.get(), mSourceFormat.get()) != AMEDIA_OK) {
LOG(ERROR) << "Unable to copy source format";
return AMEDIA_ERROR_INVALID_PARAMETER;
}
// Request decoder to convert HDR content to SDR.
const bool sourceIsHdr = VideoIsHdr(mSourceFormat.get());
if (sourceIsHdr) {
AMediaFormat_setInt32(decoderFormat.get(),
TBD_AMEDIACODEC_PARAMETER_KEY_COLOR_TRANSFER_REQUEST,
COLOR_TRANSFER_SDR_VIDEO);
}
// Prevent decoder from overwriting frames that the encoder has not yet consumed.
AMediaFormat_setInt32(decoderFormat.get(), TBD_AMEDIACODEC_PARAMETER_KEY_ALLOW_FRAME_DROP, 0);
// Copy over configurations that apply to both encoder and decoder.
static const std::vector<EntryCopier> kEncoderEntriesToCopy{
ENTRY_COPIER2(AMEDIAFORMAT_KEY_OPERATING_RATE, Float, Int32),
ENTRY_COPIER(AMEDIAFORMAT_KEY_PRIORITY, Int32),
ENTRY_COPIER(TBD_AMEDIACODEC_PARAMETER_KEY_BACKGROUND_MODE, Int32),
};
CopyFormatEntries(mDestinationFormat.get(), decoderFormat.get(), kEncoderEntriesToCopy);
LOG(INFO) << "Configuring decoder with: " << AMediaFormat_toString(decoderFormat.get());
status = AMediaCodec_configure(mDecoder, decoderFormat.get(), mSurface, NULL /* crypto */,
0 /* flags */);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to configure video decoder: " << status;
return status;
}
if (sourceIsHdr) {
bool supported = false;
AMediaFormat* inputFormat = AMediaCodec_getInputFormat(mDecoder);
if (inputFormat != nullptr) {
int32_t transferFunc;
supported = AMediaFormat_getInt32(inputFormat,
TBD_AMEDIACODEC_PARAMETER_KEY_COLOR_TRANSFER_REQUEST,
&transferFunc) &&
transferFunc == COLOR_TRANSFER_SDR_VIDEO;
AMediaFormat_delete(inputFormat);
}
if (!supported) {
LOG(ERROR) << "HDR to SDR conversion unsupported by the codec";
return AMEDIA_ERROR_UNSUPPORTED;
}
}
// Configure codecs to run in async mode.
AMediaCodecOnAsyncNotifyCallback asyncCodecCallbacks = {
.onAsyncInputAvailable = AsyncCodecCallbackDispatch::onAsyncInputAvailable,
.onAsyncOutputAvailable = AsyncCodecCallbackDispatch::onAsyncOutputAvailable,
.onAsyncFormatChanged = AsyncCodecCallbackDispatch::onAsyncFormatChanged,
.onAsyncError = AsyncCodecCallbackDispatch::onAsyncError};
// Note: The decoder does not need its own wrapper because its lifetime is tied to the
// transcoder. But the same callbacks are reused for decoder and encoder so we pass the encoder
// wrapper as userdata here but never read the codec from it in the callback.
status = AMediaCodec_setAsyncNotifyCallback(mDecoder, asyncCodecCallbacks, mEncoder.get());
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to set decoder to async mode: " << status;
return status;
}
status = AMediaCodec_setAsyncNotifyCallback(mEncoder->getCodec(), asyncCodecCallbacks,
mEncoder.get());
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to set encoder to async mode: " << status;
return status;
}
return AMEDIA_OK;
}
void VideoTrackTranscoder::enqueueInputSample(int32_t bufferIndex) {
media_status_t status = AMEDIA_OK;
if (mEosFromSource) {
return;
}
status = mMediaSampleReader->getSampleInfoForTrack(mTrackIndex, &mSampleInfo);
if (status != AMEDIA_OK && status != AMEDIA_ERROR_END_OF_STREAM) {
LOG(ERROR) << "Error getting next sample info: " << status;
mStatus = status;
return;
}
const bool endOfStream = (status == AMEDIA_ERROR_END_OF_STREAM);
if (!endOfStream) {
size_t bufferSize = 0;
uint8_t* sourceBuffer = AMediaCodec_getInputBuffer(mDecoder, bufferIndex, &bufferSize);
if (sourceBuffer == nullptr) {
LOG(ERROR) << "Decoder returned a NULL input buffer.";
mStatus = AMEDIA_ERROR_UNKNOWN;
return;
} else if (bufferSize < mSampleInfo.size) {
LOG(ERROR) << "Decoder returned an input buffer that is smaller than the sample.";
mStatus = AMEDIA_ERROR_UNKNOWN;
return;
}
status = mMediaSampleReader->readSampleDataForTrack(mTrackIndex, sourceBuffer,
mSampleInfo.size);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to read next sample data. Aborting transcode.";
mStatus = status;
return;
}
if (mSampleInfo.size) {
++mInputFrameCount;
}
} else {
LOG(DEBUG) << "EOS from source.";
mEosFromSource = true;
}
status = AMediaCodec_queueInputBuffer(mDecoder, bufferIndex, 0, mSampleInfo.size,
mSampleInfo.presentationTimeUs, mSampleInfo.flags);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to queue input buffer for decode: " << status;
mStatus = status;
return;
}
}
void VideoTrackTranscoder::transferBuffer(int32_t bufferIndex, AMediaCodecBufferInfo bufferInfo) {
if (bufferIndex >= 0) {
bool needsRender = bufferInfo.size > 0;
AMediaCodec_releaseOutputBuffer(mDecoder, bufferIndex, needsRender);
}
if (bufferInfo.flags & AMEDIACODEC_BUFFER_FLAG_END_OF_STREAM) {
LOG(DEBUG) << "EOS from decoder.";
media_status_t status = AMediaCodec_signalEndOfInputStream(mEncoder->getCodec());
if (status != AMEDIA_OK) {
LOG(ERROR) << "SignalEOS on encoder returned error: " << status;
mStatus = status;
}
}
}
void VideoTrackTranscoder::dequeueOutputSample(int32_t bufferIndex,
AMediaCodecBufferInfo bufferInfo) {
if (bufferIndex >= 0) {
size_t sampleSize = 0;
uint8_t* buffer =
AMediaCodec_getOutputBuffer(mEncoder->getCodec(), bufferIndex, &sampleSize);
MediaSample::OnSampleReleasedCallback bufferReleaseCallback =
[encoder = mEncoder](MediaSample* sample) {
AMediaCodec_releaseOutputBuffer(encoder->getCodec(), sample->bufferId,
false /* render */);
};
std::shared_ptr<MediaSample> sample = MediaSample::createWithReleaseCallback(
buffer, bufferInfo.offset, bufferIndex, bufferReleaseCallback);
sample->info.size = bufferInfo.size;
sample->info.flags = bufferInfo.flags;
sample->info.presentationTimeUs = bufferInfo.presentationTimeUs;
if (bufferInfo.size > 0 && (bufferInfo.flags & SAMPLE_FLAG_CODEC_CONFIG) == 0) {
++mOutputFrameCount;
}
onOutputSampleAvailable(sample);
mLastSampleWasSync = sample->info.flags & SAMPLE_FLAG_SYNC_SAMPLE;
}
if (bufferInfo.flags & AMEDIACODEC_BUFFER_FLAG_END_OF_STREAM) {
LOG(DEBUG) << "EOS from encoder.";
mEosFromEncoder = true;
if (mInputFrameCount != mOutputFrameCount) {
LOG(WARNING) << "Input / Output frame count mismatch: " << mInputFrameCount << " vs "
<< mOutputFrameCount;
if (mInputFrameCount > 0 && mOutputFrameCount == 0) {
LOG(ERROR) << "Encoder did not produce any output frames.";
mStatus = AMEDIA_ERROR_UNKNOWN;
}
}
}
}
void VideoTrackTranscoder::updateTrackFormat(AMediaFormat* outputFormat, bool fromDecoder) {
if (fromDecoder) {
static const std::vector<AMediaFormatUtils::EntryCopier> kValuesToCopy{
ENTRY_COPIER(AMEDIAFORMAT_KEY_COLOR_RANGE, Int32),
ENTRY_COPIER(AMEDIAFORMAT_KEY_COLOR_STANDARD, Int32),
ENTRY_COPIER(AMEDIAFORMAT_KEY_COLOR_TRANSFER, Int32),
};
AMediaFormat* params = AMediaFormat_new();
if (params != nullptr) {
AMediaFormatUtils::CopyFormatEntries(outputFormat, params, kValuesToCopy);
if (AMediaCodec_setParameters(mEncoder->getCodec(), params) != AMEDIA_OK) {
LOG(WARNING) << "Unable to update encoder with color information";
}
AMediaFormat_delete(params);
}
return;
}
if (mActualOutputFormat != nullptr) {
LOG(WARNING) << "Ignoring duplicate format change.";
return;
}
AMediaFormat* formatCopy = AMediaFormat_new();
if (!formatCopy || AMediaFormat_copy(formatCopy, outputFormat) != AMEDIA_OK) {
LOG(ERROR) << "Unable to copy outputFormat";
AMediaFormat_delete(formatCopy);
mStatus = AMEDIA_ERROR_INVALID_PARAMETER;
return;
}
// Generate the actual track format for muxer based on the encoder output format,
// since many vital information comes in the encoder format (eg. CSD).
// Transfer necessary fields from the user-configured track format (derived from
// source track format and user transcoding request) where needed.
// Transfer SAR settings:
// If mDestinationFormat has SAR set, it means the original source has SAR specified
// at container level. This is supposed to override any SAR settings in the bitstream,
// thus should always be transferred to the container of the transcoded file.
int32_t sarWidth, sarHeight;
if (AMediaFormat_getInt32(mSourceFormat.get(), AMEDIAFORMAT_KEY_SAR_WIDTH, &sarWidth) &&
(sarWidth > 0) &&
AMediaFormat_getInt32(mSourceFormat.get(), AMEDIAFORMAT_KEY_SAR_HEIGHT, &sarHeight) &&
(sarHeight > 0)) {
AMediaFormat_setInt32(formatCopy, AMEDIAFORMAT_KEY_SAR_WIDTH, sarWidth);
AMediaFormat_setInt32(formatCopy, AMEDIAFORMAT_KEY_SAR_HEIGHT, sarHeight);
}
// Transfer DAR settings.
int32_t displayWidth, displayHeight;
if (AMediaFormat_getInt32(mSourceFormat.get(), AMEDIAFORMAT_KEY_DISPLAY_WIDTH, &displayWidth) &&
(displayWidth > 0) &&
AMediaFormat_getInt32(mSourceFormat.get(), AMEDIAFORMAT_KEY_DISPLAY_HEIGHT,
&displayHeight) &&
(displayHeight > 0)) {
AMediaFormat_setInt32(formatCopy, AMEDIAFORMAT_KEY_DISPLAY_WIDTH, displayWidth);
AMediaFormat_setInt32(formatCopy, AMEDIAFORMAT_KEY_DISPLAY_HEIGHT, displayHeight);
}
// Transfer rotation settings.
// Note that muxer itself doesn't take rotation from the track format. It requires
// AMediaMuxer_setOrientationHint to set the rotation. Here we pass the rotation to
// MediaSampleWriter using the track format. MediaSampleWriter will then call
// AMediaMuxer_setOrientationHint as needed.
int32_t rotation;
if (AMediaFormat_getInt32(mSourceFormat.get(), AMEDIAFORMAT_KEY_ROTATION, &rotation) &&
(rotation != 0)) {
AMediaFormat_setInt32(formatCopy, AMEDIAFORMAT_KEY_ROTATION, rotation);
}
// Transfer track duration.
// Preserve the source track duration by sending it to MediaSampleWriter.
int64_t durationUs;
if (AMediaFormat_getInt64(mSourceFormat.get(), AMEDIAFORMAT_KEY_DURATION, &durationUs) &&
durationUs > 0) {
AMediaFormat_setInt64(formatCopy, AMEDIAFORMAT_KEY_DURATION, durationUs);
}
// TODO: transfer other fields as required.
mActualOutputFormat = std::shared_ptr<AMediaFormat>(formatCopy, &AMediaFormat_delete);
LOG(INFO) << "Actual output format: " << AMediaFormat_toString(formatCopy);
notifyTrackFormatAvailable();
}
media_status_t VideoTrackTranscoder::runTranscodeLoop(bool* stopped) {
prctl(PR_SET_NAME, (unsigned long)"VideTranscodTrd", 0, 0, 0);
// Push start decoder and encoder as two messages, so that these are subject to the
// stop request as well. If the session is cancelled (or paused) immediately after start,
// we don't need to waste time start then stop the codecs.
mCodecMessageQueue.push([this] {
media_status_t status = AMediaCodec_start(mDecoder);
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to start video decoder: " << status;
mStatus = status;
}
});
mCodecMessageQueue.push([this] {
media_status_t status = AMediaCodec_start(mEncoder->getCodec());
if (status != AMEDIA_OK) {
LOG(ERROR) << "Unable to start video encoder: " << status;
mStatus = status;
}
mEncoder->setStarted();
});
// Process codec events until EOS is reached, transcoding is stopped or an error occurs.
while (mStopRequest != STOP_NOW && !mEosFromEncoder && mStatus == AMEDIA_OK) {
std::function<void()> message = mCodecMessageQueue.pop();
message();
if (mStopRequest == STOP_ON_SYNC && mLastSampleWasSync) {
break;
}
}
mCodecMessageQueue.abort();
AMediaCodec_stop(mDecoder);
// Signal if transcoding was stopped before it finished.
if (mStopRequest != NONE && !mEosFromEncoder && mStatus == AMEDIA_OK) {
*stopped = true;
}
return mStatus;
}
void VideoTrackTranscoder::abortTranscodeLoop() {
if (mStopRequest == STOP_NOW) {
// Wake up transcoder thread.
mCodecMessageQueue.push([] {}, true /* front */);
}
}
std::shared_ptr<AMediaFormat> VideoTrackTranscoder::getOutputFormat() const {
return mActualOutputFormat;
}
} // namespace android