blob: 2e7043f475730881c6c3a4b6c236122869e440cc [file] [log] [blame]
/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//#define LOG_NDEBUG 0
#include <utils/Log.h>
#include <utils/misc.h>
//#include "OMX_VideoExt.h"
#define DEBUG 0
#if DEBUG
#define DDD(...) ALOGD(__VA_ARGS__)
#else
#define DDD(...) ((void)0)
#endif
#include "GoldfishVPX.h"
#include <media/stagefright/foundation/ADebug.h>
#include <media/stagefright/MediaDefs.h>
#include <OMX_VideoExt.h>
#include <inttypes.h>
#include <nativebase/nativebase.h>
#include <android/hardware/graphics/allocator/3.0/IAllocator.h>
#include <android/hardware/graphics/mapper/3.0/IMapper.h>
#include <hidl/LegacySupport.h>
using ::android::hardware::graphics::common::V1_0::BufferUsage;
using ::android::hardware::graphics::common::V1_2::PixelFormat;
namespace android {
// Only need to declare the highest supported profile and level here.
static const CodecProfileLevel kVP9ProfileLevels[] = {
{ OMX_VIDEO_VP9Profile0, OMX_VIDEO_VP9Level5 },
{ OMX_VIDEO_VP9Profile2, OMX_VIDEO_VP9Level5 },
{ OMX_VIDEO_VP9Profile2HDR, OMX_VIDEO_VP9Level5 },
{ OMX_VIDEO_VP9Profile2HDR10Plus, OMX_VIDEO_VP9Level5 },
};
GoldfishVPX::GoldfishVPX(const char* name,
const char* componentRole,
OMX_VIDEO_CODINGTYPE codingType,
const OMX_CALLBACKTYPE* callbacks,
OMX_PTR appData,
OMX_COMPONENTTYPE** component,
RenderMode renderMode)
: GoldfishVideoDecoderOMXComponent(
name,
componentRole,
codingType,
codingType == OMX_VIDEO_CodingVP8 ? NULL : kVP9ProfileLevels,
codingType == OMX_VIDEO_CodingVP8 ? 0 : NELEM(kVP9ProfileLevels),
320 /* width */,
240 /* height */,
callbacks,
appData,
component),
mMode(codingType == OMX_VIDEO_CodingVP8 ? MODE_VP8 : MODE_VP9),
mRenderMode(renderMode),
mEOSStatus(INPUT_DATA_AVAILABLE),
mCtx(NULL),
mFrameParallelMode(false),
mTimeStampIdx(0),
mImg(NULL) {
// arbitrary from avc/hevc as vpx does not specify a min compression ratio
const size_t kMinCompressionRatio = mMode == MODE_VP8 ? 2 : 4;
const char* mime = mMode == MODE_VP8 ? MEDIA_MIMETYPE_VIDEO_VP8
: MEDIA_MIMETYPE_VIDEO_VP9;
const size_t kMaxOutputBufferSize = 3840 * 2160 * 3 / 2; // 4k
initPorts(kNumBuffers,
kMaxOutputBufferSize / kMinCompressionRatio /* inputBufferSize */,
kNumBuffers, mime, kMinCompressionRatio);
ALOGI("calling constructor of GoldfishVPX");
// wait till later
// CHECK_EQ(initDecoder(), (status_t)OK);
}
GoldfishVPX::~GoldfishVPX() {
ALOGI("calling destructor of GoldfishVPX");
destroyDecoder();
}
bool GoldfishVPX::supportDescribeHdrStaticInfo() {
return true;
}
bool GoldfishVPX::supportDescribeHdr10PlusInfo() {
return true;
}
status_t GoldfishVPX::initDecoder() {
mCtx = new vpx_codec_ctx_t;
mCtx->vpversion = mMode == MODE_VP8 ? 8 : 9;
mCtx->version = mEnableAndroidNativeBuffers ? 200 : 100;
int vpx_err = 0;
if ((vpx_err = vpx_codec_dec_init(mCtx))) {
ALOGE("vpx decoder failed to initialize. (%d)", vpx_err);
return UNKNOWN_ERROR;
}
ALOGI("calling init GoldfishVPX ctx %p", mCtx);
return OK;
}
status_t GoldfishVPX::destroyDecoder() {
if (mCtx) {
ALOGI("calling destroying GoldfishVPX ctx %p", mCtx);
vpx_codec_destroy(mCtx);
delete mCtx;
mCtx = NULL;
}
return OK;
}
void GoldfishVPX::setup_ctx_parameters(vpx_codec_ctx_t* ctx,
int hostColorBufferId) {
ctx->width = mWidth;
ctx->height = mHeight;
ctx->hostColorBufferId = hostColorBufferId;
ctx->outputBufferWidth = outputBufferWidth();
ctx->outputBufferHeight = outputBufferHeight();
OMX_PARAM_PORTDEFINITIONTYPE *outDef = &editPortInfo(kOutputPortIndex)->mDef;
int32_t bpp = (outDef->format.video.eColorFormat == OMX_COLOR_FormatYUV420Planar16) ? 2 : 1;
ctx->bpp = bpp;
}
bool GoldfishVPX::outputBuffers(bool flushDecoder, bool display, bool eos, bool *portWillReset) {
List<BufferInfo *> &outQueue = getPortQueue(1);
BufferInfo *outInfo = NULL;
OMX_BUFFERHEADERTYPE *outHeader = NULL;
DDD("%s %d", __func__, __LINE__);
if (flushDecoder && mFrameParallelMode) {
// Flush decoder by passing NULL data ptr and 0 size.
// Ideally, this should never fail.
if (vpx_codec_flush(mCtx)) {
ALOGE("Failed to flush on2 decoder.");
return false;
}
}
if (!display) {
if (!flushDecoder) {
ALOGE("Invalid operation.");
return false;
}
// Drop all the decoded frames in decoder.
// TODO: move this to host, with something like
// vpx_codec_drop_all_frames(mCtx);
setup_ctx_parameters(mCtx);
while ((mImg = vpx_codec_get_frame(mCtx))) {
}
return true;
}
while (!outQueue.empty()) {
DDD("%s %d", __func__, __LINE__);
outInfo = *outQueue.begin();
outHeader = outInfo->mHeader;
if (mImg == NULL) {
setup_ctx_parameters(mCtx, getHostColorBufferId(outHeader));
mImg = vpx_codec_get_frame(mCtx);
if (mImg == NULL) {
break;
}
}
uint32_t width = mImg->d_w;
uint32_t height = mImg->d_h;
CHECK(mImg->fmt == VPX_IMG_FMT_I420 || mImg->fmt == VPX_IMG_FMT_I42016);
OMX_COLOR_FORMATTYPE outputColorFormat = OMX_COLOR_FormatYUV420Planar;
int32_t bpp = 1;
if (mImg->fmt == VPX_IMG_FMT_I42016) {
outputColorFormat = OMX_COLOR_FormatYUV420Planar16;
bpp = 2;
}
handlePortSettingsChange(portWillReset, width, height, outputColorFormat);
if (*portWillReset) {
return true;
}
outHeader->nOffset = 0;
outHeader->nFlags = 0;
outHeader->nFilledLen = (outputBufferWidth() * outputBufferHeight() * bpp * 3) / 2;
PrivInfo *privInfo = (PrivInfo *)mImg->user_priv;
outHeader->nTimeStamp = privInfo->mTimeStamp;
if (privInfo->mHdr10PlusInfo != nullptr) {
queueOutputFrameConfig(privInfo->mHdr10PlusInfo);
}
if (outputBufferSafe(outHeader) &&
getHostColorBufferId(outHeader) < 0) {
uint8_t *dst = outHeader->pBuffer;
memcpy(dst, mCtx->dst, outHeader->nFilledLen);
} else {
// outHeader->nFilledLen = 0;
}
mImg = NULL;
outInfo->mOwnedByUs = false;
outQueue.erase(outQueue.begin());
outInfo = NULL;
notifyFillBufferDone(outHeader);
outHeader = NULL;
}
if (!eos) {
return true;
}
if (!outQueue.empty()) {
outInfo = *outQueue.begin();
outQueue.erase(outQueue.begin());
outHeader = outInfo->mHeader;
outHeader->nTimeStamp = 0;
outHeader->nFilledLen = 0;
outHeader->nFlags = OMX_BUFFERFLAG_EOS;
outInfo->mOwnedByUs = false;
notifyFillBufferDone(outHeader);
mEOSStatus = OUTPUT_FRAMES_FLUSHED;
}
return true;
}
bool GoldfishVPX::outputBufferSafe(OMX_BUFFERHEADERTYPE *outHeader) {
DDD("%s %d", __func__, __LINE__);
uint32_t width = outputBufferWidth();
uint32_t height = outputBufferHeight();
uint64_t nFilledLen = width;
nFilledLen *= height;
if (nFilledLen > UINT32_MAX / 3) {
ALOGE("b/29421675, nFilledLen overflow %llu w %u h %u",
(unsigned long long)nFilledLen, width, height);
android_errorWriteLog(0x534e4554, "29421675");
return false;
} else if (outHeader->nAllocLen < outHeader->nFilledLen) {
ALOGE("b/27597103, buffer too small");
android_errorWriteLog(0x534e4554, "27597103");
return false;
}
return true;
}
void GoldfishVPX::onQueueFilled(OMX_U32 /* portIndex */) {
DDD("%s %d", __func__, __LINE__);
if (mOutputPortSettingsChange != NONE || mEOSStatus == OUTPUT_FRAMES_FLUSHED) {
return;
}
if (mCtx == nullptr) {
if (OK != initDecoder()) {
ALOGE("Failed to initialize decoder");
notify(OMX_EventError, OMX_ErrorUnsupportedSetting, 0, NULL);
return;
}
}
List<BufferInfo *> &inQueue = getPortQueue(0);
List<BufferInfo *> &outQueue = getPortQueue(1);
bool EOSseen = false;
bool portWillReset = false;
while ((mEOSStatus == INPUT_EOS_SEEN || !inQueue.empty())
&& !outQueue.empty()) {
// Output the pending frames that left from last port reset or decoder flush.
if (mEOSStatus == INPUT_EOS_SEEN || mImg != NULL) {
if (!outputBuffers(
mEOSStatus == INPUT_EOS_SEEN, true /* display */,
mEOSStatus == INPUT_EOS_SEEN, &portWillReset)) {
ALOGE("on2 decoder failed to output frame.");
notify(OMX_EventError, OMX_ErrorUndefined, 0, NULL);
return;
}
if (portWillReset || mEOSStatus == OUTPUT_FRAMES_FLUSHED ||
mEOSStatus == INPUT_EOS_SEEN) {
return;
}
// Continue as outQueue may be empty now.
continue;
}
BufferInfo *inInfo = *inQueue.begin();
OMX_BUFFERHEADERTYPE *inHeader = inInfo->mHeader;
// Software VP9 Decoder does not need the Codec Specific Data (CSD)
// (specified in http://www.webmproject.org/vp9/profiles/). Ignore it if
// it was passed.
if (inHeader->nFlags & OMX_BUFFERFLAG_CODECCONFIG) {
// Only ignore CSD buffer for VP9.
if (mMode == MODE_VP9) {
inQueue.erase(inQueue.begin());
inInfo->mOwnedByUs = false;
notifyEmptyBufferDone(inHeader);
continue;
} else {
// Tolerate the CSD buffer for VP8. This is a workaround
// for b/28689536.
ALOGW("WARNING: Got CSD buffer for VP8.");
}
}
mPrivInfo[mTimeStampIdx].mTimeStamp = inHeader->nTimeStamp;
if (inInfo->mFrameConfig) {
mPrivInfo[mTimeStampIdx].mHdr10PlusInfo = dequeueInputFrameConfig();
} else {
mPrivInfo[mTimeStampIdx].mHdr10PlusInfo.clear();
}
if (inHeader->nFlags & OMX_BUFFERFLAG_EOS) {
mEOSStatus = INPUT_EOS_SEEN;
EOSseen = true;
}
if (inHeader->nFilledLen > 0) {
int err = vpx_codec_decode(mCtx, inHeader->pBuffer + inHeader->nOffset,
inHeader->nFilledLen, &mPrivInfo[mTimeStampIdx], 0);
if (err == VPX_CODEC_OK) {
inInfo->mOwnedByUs = false;
inQueue.erase(inQueue.begin());
inInfo = NULL;
notifyEmptyBufferDone(inHeader);
inHeader = NULL;
} else {
ALOGE("on2 decoder failed to decode frame. err: %d", err);
notify(OMX_EventError, OMX_ErrorUndefined, 0, NULL);
return;
}
}
mTimeStampIdx = (mTimeStampIdx + 1) % kNumBuffers;
if (!outputBuffers(
EOSseen /* flushDecoder */, true /* display */, EOSseen, &portWillReset)) {
ALOGE("on2 decoder failed to output frame.");
notify(OMX_EventError, OMX_ErrorUndefined, 0, NULL);
return;
}
if (portWillReset) {
return;
}
}
}
void GoldfishVPX::onPortFlushCompleted(OMX_U32 portIndex) {
DDD("%s %d", __func__, __LINE__);
if (portIndex == kInputPortIndex) {
bool portWillReset = false;
if (!outputBuffers(
true /* flushDecoder */, false /* display */, false /* eos */, &portWillReset)) {
ALOGE("Failed to flush decoder.");
notify(OMX_EventError, OMX_ErrorUndefined, 0, NULL);
return;
}
mEOSStatus = INPUT_DATA_AVAILABLE;
}
}
void GoldfishVPX::onReset() {
DDD("%s %d", __func__, __LINE__);
bool portWillReset = false;
if (!outputBuffers(
true /* flushDecoder */, false /* display */, false /* eos */, &portWillReset)) {
ALOGW("Failed to flush decoder. Try to hard reset decoder");
destroyDecoder();
initDecoder();
}
mEOSStatus = INPUT_DATA_AVAILABLE;
}
int GoldfishVPX::getHostColorBufferId(void* header) {
DDD("%s %d", __func__, __LINE__);
if (mNWBuffers.find(header) == mNWBuffers.end()) {
DDD("cannot find color buffer for header %p", header);
return -1;
}
sp<ANativeWindowBuffer> nBuf = mNWBuffers[header];
cb_handle_t* handle = (cb_handle_t*)nBuf->handle;
DDD("found color buffer for header %p --> %d", header, handle->hostHandle);
return handle->hostHandle;
}
OMX_ERRORTYPE GoldfishVPX::internalGetParameter(OMX_INDEXTYPE index,
OMX_PTR params) {
const int32_t indexFull = index;
switch (indexFull) {
case kGetAndroidNativeBufferUsageIndex: {
DDD("calling kGetAndroidNativeBufferUsageIndex");
GetAndroidNativeBufferUsageParams* nativeBuffersUsage =
(GetAndroidNativeBufferUsageParams*)params;
nativeBuffersUsage->nUsage =
(unsigned int)(BufferUsage::GPU_DATA_BUFFER);
return OMX_ErrorNone;
}
default:
return GoldfishVideoDecoderOMXComponent::internalGetParameter(
index, params);
}
}
OMX_ERRORTYPE GoldfishVPX::internalSetParameter(OMX_INDEXTYPE index,
const OMX_PTR params) {
// Include extension index OMX_INDEXEXTTYPE.
const int32_t indexFull = index;
switch (indexFull) {
case kEnableAndroidNativeBuffersIndex: {
DDD("calling kEnableAndroidNativeBuffersIndex");
EnableAndroidNativeBuffersParams* enableNativeBuffers =
(EnableAndroidNativeBuffersParams*)params;
if (enableNativeBuffers) {
mEnableAndroidNativeBuffers = enableNativeBuffers->enable;
if (mEnableAndroidNativeBuffers == false) {
mNWBuffers.clear();
DDD("disabled kEnableAndroidNativeBuffersIndex");
} else {
DDD("enabled kEnableAndroidNativeBuffersIndex");
}
}
return OMX_ErrorNone;
}
case kUseAndroidNativeBufferIndex: {
if (mEnableAndroidNativeBuffers == false) {
ALOGE("Error: not enabled Android Native Buffers");
return OMX_ErrorBadParameter;
}
UseAndroidNativeBufferParams* use_buffer_params =
(UseAndroidNativeBufferParams*)params;
if (use_buffer_params) {
sp<ANativeWindowBuffer> nBuf = use_buffer_params->nativeBuffer;
cb_handle_t* handle = (cb_handle_t*)nBuf->handle;
void* dst = NULL;
DDD("kUseAndroidNativeBufferIndex with handle %p host color "
"handle %d "
"calling usebuffer",
handle, handle->hostHandle);
useBufferCallerLockedAlready(use_buffer_params->bufferHeader,
use_buffer_params->nPortIndex,
use_buffer_params->pAppPrivate,
handle->allocatedSize(),
(OMX_U8*)dst);
mNWBuffers[*(use_buffer_params->bufferHeader)] =
use_buffer_params->nativeBuffer;
;
}
return OMX_ErrorNone;
}
default:
return GoldfishVideoDecoderOMXComponent::internalSetParameter(
index, params);
}
}
OMX_ERRORTYPE GoldfishVPX::getExtensionIndex(const char* name,
OMX_INDEXTYPE* index) {
if (mRenderMode == RenderMode::RENDER_BY_HOST_GPU) {
if (!strcmp(name,
"OMX.google.android.index.enableAndroidNativeBuffers")) {
DDD("calling getExtensionIndex for enable ANB");
*(int32_t*)index = kEnableAndroidNativeBuffersIndex;
return OMX_ErrorNone;
} else if (!strcmp(name,
"OMX.google.android.index.useAndroidNativeBuffer")) {
*(int32_t*)index = kUseAndroidNativeBufferIndex;
return OMX_ErrorNone;
} else if (!strcmp(name,
"OMX.google.android.index."
"getAndroidNativeBufferUsage")) {
*(int32_t*)index = kGetAndroidNativeBufferUsageIndex;
return OMX_ErrorNone;
}
}
return GoldfishVideoDecoderOMXComponent::getExtensionIndex(name, index);
}
} // namespace android
android::GoldfishOMXComponent *createGoldfishOMXComponent(
const char *name, const OMX_CALLBACKTYPE *callbacks,
OMX_PTR appData, OMX_COMPONENTTYPE **component) {
DDD("%s %d", __func__, __LINE__);
// only support vp9 to use host hardware decoder, for now
if (!strncmp("OMX.android.goldfish.vp9.decoder", name, 32)) {
return new android::GoldfishVPX(
name, "video_decoder.vp9", OMX_VIDEO_CodingVP9, callbacks,
appData, component, RenderMode::RENDER_BY_HOST_GPU);
}
if (!strncmp("OMX.android.goldfish.vp8.decoder", name, 32)) {
return new android::GoldfishVPX(
name, "video_decoder.vp8", OMX_VIDEO_CodingVP8, callbacks,
appData, component, RenderMode::RENDER_BY_HOST_GPU);
}
if (!strncmp("OMX.google.goldfish.vp9.decoder", name, 30)) {
return new android::GoldfishVPX(
name, "video_decoder.vp9", OMX_VIDEO_CodingVP9, callbacks,
appData, component, RenderMode::RENDER_BY_GUEST_CPU);
}
if (!strncmp("OMX.google.goldfish.vp8.decoder", name, 30)) {
return new android::GoldfishVPX(
name, "video_decoder.vp8", OMX_VIDEO_CodingVP8, callbacks,
appData, component, RenderMode::RENDER_BY_GUEST_CPU);
}
{ CHECK(!"Unknown component"); }
return NULL;
}