blob: a3f55da7116afd0b976aec1b23c6917e0578db50 [file] [log] [blame] [edit]
/*
* Copyright 2015 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//#define LOG_NDEBUG 0
#define LOG_TAG "MediaSync"
#include <inttypes.h>
#include <gui/BufferQueue.h>
#include <gui/IGraphicBufferConsumer.h>
#include <gui/IGraphicBufferProducer.h>
#include <media/AudioTrack.h>
#include <media/stagefright/MediaClock.h>
#include <media/stagefright/MediaSync.h>
#include <media/stagefright/VideoFrameScheduler.h>
#include <media/stagefright/foundation/ADebug.h>
#include <media/stagefright/foundation/ALooper.h>
#include <media/stagefright/foundation/AMessage.h>
#include <ui/GraphicBuffer.h>
#include <system/window.h>
// Maximum late time allowed for a video frame to be rendered. When a video
// frame arrives later than this number, it will be discarded without rendering.
static const int64_t kMaxAllowedVideoLateTimeUs = 40000LL;
namespace android {
// static
sp<MediaSync> MediaSync::create() {
sp<MediaSync> sync = new MediaSync();
sync->mLooper->registerHandler(sync);
return sync;
}
MediaSync::MediaSync()
: mIsAbandoned(false),
mMutex(),
mReleaseCondition(),
mNumOutstandingBuffers(0),
mUsageFlagsFromOutput(0),
mMaxAcquiredBufferCount(1),
mReturnPendingInputFrame(false),
mNativeSampleRateInHz(0),
mNumFramesWritten(0),
mHasAudio(false),
mNextBufferItemMediaUs(-1),
mPlaybackRate(0.0) {
mMediaClock = new MediaClock;
mMediaClock->init();
// initialize settings
mPlaybackSettings = AUDIO_PLAYBACK_RATE_DEFAULT;
mPlaybackSettings.mSpeed = mPlaybackRate;
mLooper = new ALooper;
mLooper->setName("MediaSync");
mLooper->start(false, false, ANDROID_PRIORITY_AUDIO);
}
MediaSync::~MediaSync() {
if (mInput != NULL) {
mInput->consumerDisconnect();
}
if (mOutput != NULL) {
mOutput->disconnect(NATIVE_WINDOW_API_MEDIA);
}
if (mLooper != NULL) {
mLooper->unregisterHandler(id());
mLooper->stop();
}
}
status_t MediaSync::setSurface(const sp<IGraphicBufferProducer> &output) {
Mutex::Autolock lock(mMutex);
if (output == mOutput) {
return NO_ERROR; // same output surface.
}
if (output == NULL && mSyncSettings.mSource == AVSYNC_SOURCE_VSYNC) {
ALOGE("setSurface: output surface is used as sync source and cannot be removed.");
return INVALID_OPERATION;
}
if (output != NULL) {
int newUsage = 0;
output->query(NATIVE_WINDOW_CONSUMER_USAGE_BITS, &newUsage);
// Check usage flags only when current output surface has been used to create input surface.
if (mOutput != NULL && mInput != NULL) {
int ignoredFlags = (GRALLOC_USAGE_HW_TEXTURE | GRALLOC_USAGE_HW_COMPOSER
| GRALLOC_USAGE_EXTERNAL_DISP);
// New output surface is not allowed to add new usage flag except ignored ones.
if ((newUsage & ~(mUsageFlagsFromOutput | ignoredFlags)) != 0) {
ALOGE("setSurface: new output surface has new usage flag not used by current one.");
return BAD_VALUE;
}
}
// Try to connect to new output surface. If failed, current output surface will not
// be changed.
IGraphicBufferProducer::QueueBufferOutput queueBufferOutput;
sp<OutputListener> listener(new OutputListener(this, output));
IInterface::asBinder(output)->linkToDeath(listener);
status_t status =
output->connect(listener,
NATIVE_WINDOW_API_MEDIA,
true /* producerControlledByApp */,
&queueBufferOutput);
if (status != NO_ERROR) {
ALOGE("setSurface: failed to connect (%d)", status);
return status;
}
if (mFrameScheduler == NULL) {
mFrameScheduler = new VideoFrameScheduler();
mFrameScheduler->init();
}
}
if (mOutput != NULL) {
mOutput->disconnect(NATIVE_WINDOW_API_MEDIA);
while (!mBuffersSentToOutput.isEmpty()) {
returnBufferToInput_l(mBuffersSentToOutput.valueAt(0), Fence::NO_FENCE);
mBuffersSentToOutput.removeItemsAt(0);
}
}
mOutput = output;
return NO_ERROR;
}
// |audioTrack| is used only for querying information.
status_t MediaSync::setAudioTrack(const sp<AudioTrack> &audioTrack) {
Mutex::Autolock lock(mMutex);
// TODO: support audio track change.
if (mAudioTrack != NULL) {
ALOGE("setAudioTrack: audioTrack has already been configured.");
return INVALID_OPERATION;
}
if (audioTrack == NULL && mSyncSettings.mSource == AVSYNC_SOURCE_AUDIO) {
ALOGE("setAudioTrack: audioTrack is used as sync source and cannot be removed.");
return INVALID_OPERATION;
}
if (audioTrack != NULL) {
// check if audio track supports the playback settings
if (mPlaybackSettings.mSpeed != 0.f
&& audioTrack->setPlaybackRate(mPlaybackSettings) != OK) {
ALOGE("playback settings are not supported by the audio track");
return INVALID_OPERATION;
}
uint32_t nativeSampleRateInHz = audioTrack->getOriginalSampleRate();
if (nativeSampleRateInHz <= 0) {
ALOGE("setAudioTrack: native sample rate should be positive.");
return BAD_VALUE;
}
mAudioTrack = audioTrack;
mNativeSampleRateInHz = nativeSampleRateInHz;
(void)setPlaybackSettings_l(mPlaybackSettings);
}
else {
mAudioTrack = NULL;
mNativeSampleRateInHz = 0;
}
// potentially resync to new source
resync_l();
return OK;
}
status_t MediaSync::createInputSurface(
sp<IGraphicBufferProducer> *outBufferProducer) {
if (outBufferProducer == NULL) {
return BAD_VALUE;
}
Mutex::Autolock lock(mMutex);
if (mOutput == NULL) {
return NO_INIT;
}
if (mInput != NULL) {
return INVALID_OPERATION;
}
sp<IGraphicBufferProducer> bufferProducer;
sp<IGraphicBufferConsumer> bufferConsumer;
BufferQueue::createBufferQueue(&bufferProducer, &bufferConsumer);
sp<InputListener> listener(new InputListener(this));
IInterface::asBinder(bufferConsumer)->linkToDeath(listener);
status_t status =
bufferConsumer->consumerConnect(listener, false /* controlledByApp */);
if (status == NO_ERROR) {
bufferConsumer->setConsumerName(String8("MediaSync"));
// propagate usage bits from output surface
mUsageFlagsFromOutput = 0;
mOutput->query(NATIVE_WINDOW_CONSUMER_USAGE_BITS, &mUsageFlagsFromOutput);
bufferConsumer->setConsumerUsageBits(mUsageFlagsFromOutput);
*outBufferProducer = bufferProducer;
mInput = bufferConsumer;
// set undequeued buffer count
int minUndequeuedBuffers;
mOutput->query(NATIVE_WINDOW_MIN_UNDEQUEUED_BUFFERS, &minUndequeuedBuffers);
mMaxAcquiredBufferCount = minUndequeuedBuffers;
bufferConsumer->setMaxAcquiredBufferCount(mMaxAcquiredBufferCount);
}
return status;
}
void MediaSync::resync_l() {
AVSyncSource src = mSyncSettings.mSource;
if (src == AVSYNC_SOURCE_DEFAULT) {
if (mAudioTrack != NULL) {
src = AVSYNC_SOURCE_AUDIO;
} else {
src = AVSYNC_SOURCE_SYSTEM_CLOCK;
}
}
// TODO: resync ourselves to the current clock (e.g. on sync source change)
updatePlaybackRate_l(mPlaybackRate);
}
void MediaSync::updatePlaybackRate_l(float rate) {
if (rate > mPlaybackRate) {
mNextBufferItemMediaUs = -1;
}
mPlaybackRate = rate;
// TODO: update frame scheduler with this info
mMediaClock->setPlaybackRate(rate);
onDrainVideo_l();
}
sp<const MediaClock> MediaSync::getMediaClock() {
return mMediaClock;
}
status_t MediaSync::getPlayTimeForPendingAudioFrames(int64_t *outTimeUs) {
Mutex::Autolock lock(mMutex);
// User should check the playback rate if it doesn't want to receive a
// huge number for play time.
if (mPlaybackRate == 0.0f) {
*outTimeUs = INT64_MAX;
return OK;
}
uint32_t numFramesPlayed = 0;
if (mAudioTrack != NULL) {
status_t res = mAudioTrack->getPosition(&numFramesPlayed);
if (res != OK) {
return res;
}
}
int64_t numPendingFrames = mNumFramesWritten - numFramesPlayed;
if (numPendingFrames < 0) {
numPendingFrames = 0;
ALOGW("getPlayTimeForPendingAudioFrames: pending frame count is negative.");
}
double timeUs = numPendingFrames * 1000000.0
/ (mNativeSampleRateInHz * (double)mPlaybackRate);
if (timeUs > (double)INT64_MAX) {
// Overflow.
*outTimeUs = INT64_MAX;
ALOGW("getPlayTimeForPendingAudioFrames: play time for pending audio frames "
"is too high, possibly due to super low playback rate(%f)", mPlaybackRate);
} else {
*outTimeUs = (int64_t)timeUs;
}
return OK;
}
status_t MediaSync::updateQueuedAudioData(
size_t sizeInBytes, int64_t presentationTimeUs) {
if (sizeInBytes == 0) {
return OK;
}
Mutex::Autolock lock(mMutex);
if (mAudioTrack == NULL) {
ALOGW("updateQueuedAudioData: audioTrack has NOT been configured.");
return INVALID_OPERATION;
}
int64_t numFrames = sizeInBytes / mAudioTrack->frameSize();
int64_t maxMediaTimeUs = presentationTimeUs
+ getDurationIfPlayedAtNativeSampleRate_l(numFrames);
int64_t nowUs = ALooper::GetNowUs();
int64_t nowMediaUs = presentationTimeUs
- getDurationIfPlayedAtNativeSampleRate_l(mNumFramesWritten)
+ getPlayedOutAudioDurationMedia_l(nowUs);
mNumFramesWritten += numFrames;
int64_t oldRealTime = -1;
if (mNextBufferItemMediaUs != -1) {
oldRealTime = getRealTime(mNextBufferItemMediaUs, nowUs);
}
mMediaClock->updateAnchor(nowMediaUs, nowUs, maxMediaTimeUs);
mHasAudio = true;
if (oldRealTime != -1) {
int64_t newRealTime = getRealTime(mNextBufferItemMediaUs, nowUs);
if (newRealTime >= oldRealTime) {
return OK;
}
}
mNextBufferItemMediaUs = -1;
onDrainVideo_l();
return OK;
}
void MediaSync::setName(const AString &name) {
Mutex::Autolock lock(mMutex);
mInput->setConsumerName(String8(name.c_str()));
}
void MediaSync::flush() {
Mutex::Autolock lock(mMutex);
if (mFrameScheduler != NULL) {
mFrameScheduler->restart();
}
while (!mBufferItems.empty()) {
BufferItem *bufferItem = &*mBufferItems.begin();
returnBufferToInput_l(bufferItem->mGraphicBuffer, bufferItem->mFence);
mBufferItems.erase(mBufferItems.begin());
}
mNextBufferItemMediaUs = -1;
mNumFramesWritten = 0;
mReturnPendingInputFrame = true;
mReleaseCondition.signal();
mMediaClock->clearAnchor();
}
status_t MediaSync::setVideoFrameRateHint(float rate) {
Mutex::Autolock lock(mMutex);
if (rate < 0.f) {
return BAD_VALUE;
}
if (mFrameScheduler != NULL) {
mFrameScheduler->init(rate);
}
return OK;
}
float MediaSync::getVideoFrameRate() {
Mutex::Autolock lock(mMutex);
if (mFrameScheduler != NULL) {
float fps = mFrameScheduler->getFrameRate();
if (fps > 0.f) {
return fps;
}
}
// we don't have or know the frame rate
return -1.f;
}
status_t MediaSync::setSyncSettings(const AVSyncSettings &syncSettings) {
// validate settings
if (syncSettings.mSource >= AVSYNC_SOURCE_MAX
|| syncSettings.mAudioAdjustMode >= AVSYNC_AUDIO_ADJUST_MODE_MAX
|| syncSettings.mTolerance < 0.f
|| syncSettings.mTolerance >= AVSYNC_TOLERANCE_MAX) {
return BAD_VALUE;
}
Mutex::Autolock lock(mMutex);
// verify that we have the sync source
switch (syncSettings.mSource) {
case AVSYNC_SOURCE_AUDIO:
if (mAudioTrack == NULL) {
ALOGE("setSyncSettings: audio sync source requires an audio track");
return BAD_VALUE;
}
break;
case AVSYNC_SOURCE_VSYNC:
if (mOutput == NULL) {
ALOGE("setSyncSettings: vsync sync source requires an output surface");
return BAD_VALUE;
}
break;
default:
break;
}
mSyncSettings = syncSettings;
resync_l();
return OK;
}
void MediaSync::getSyncSettings(AVSyncSettings *syncSettings) {
Mutex::Autolock lock(mMutex);
*syncSettings = mSyncSettings;
}
status_t MediaSync::setPlaybackSettings(const AudioPlaybackRate &rate) {
Mutex::Autolock lock(mMutex);
status_t err = setPlaybackSettings_l(rate);
if (err == OK) {
// TODO: adjust rate if using VSYNC as source
updatePlaybackRate_l(rate.mSpeed);
}
return err;
}
status_t MediaSync::setPlaybackSettings_l(const AudioPlaybackRate &rate) {
if (rate.mSpeed < 0.f || rate.mPitch < 0.f) {
// We don't validate other audio settings.
// They will be validated when/if audiotrack is set.
return BAD_VALUE;
}
if (mAudioTrack != NULL) {
if (rate.mSpeed == 0.f) {
mAudioTrack->pause();
} else {
status_t err = mAudioTrack->setPlaybackRate(rate);
if (err != OK) {
return BAD_VALUE;
}
// ignore errors
(void)mAudioTrack->start();
}
}
mPlaybackSettings = rate;
return OK;
}
void MediaSync::getPlaybackSettings(AudioPlaybackRate *rate) {
Mutex::Autolock lock(mMutex);
*rate = mPlaybackSettings;
}
int64_t MediaSync::getRealTime(int64_t mediaTimeUs, int64_t nowUs) {
int64_t realUs;
if (mMediaClock->getRealTimeFor(mediaTimeUs, &realUs) != OK) {
// If failed to get current position, e.g. due to audio clock is
// not ready, then just play out video immediately without delay.
return nowUs;
}
return realUs;
}
int64_t MediaSync::getDurationIfPlayedAtNativeSampleRate_l(int64_t numFrames) {
return (numFrames * 1000000LL / mNativeSampleRateInHz);
}
int64_t MediaSync::getPlayedOutAudioDurationMedia_l(int64_t nowUs) {
CHECK(mAudioTrack != NULL);
uint32_t numFramesPlayed;
int64_t numFramesPlayedAtUs;
AudioTimestamp ts;
status_t res = mAudioTrack->getTimestamp(ts);
if (res == OK) {
// case 1: mixing audio tracks.
numFramesPlayed = ts.mPosition;
numFramesPlayedAtUs = ts.mTime.tv_sec * 1000000LL + ts.mTime.tv_nsec / 1000;
//ALOGD("getTimestamp: OK %d %lld",
// numFramesPlayed, (long long)numFramesPlayedAtUs);
} else if (res == WOULD_BLOCK) {
// case 2: transitory state on start of a new track
numFramesPlayed = 0;
numFramesPlayedAtUs = nowUs;
//ALOGD("getTimestamp: WOULD_BLOCK %d %lld",
// numFramesPlayed, (long long)numFramesPlayedAtUs);
} else {
// case 3: transitory at new track or audio fast tracks.
res = mAudioTrack->getPosition(&numFramesPlayed);
CHECK_EQ(res, (status_t)OK);
numFramesPlayedAtUs = nowUs;
numFramesPlayedAtUs += 1000LL * mAudioTrack->latency() / 2; /* XXX */
//ALOGD("getPosition: %d %lld", numFramesPlayed, (long long)numFramesPlayedAtUs);
}
//can't be negative until 12.4 hrs, test.
//CHECK_EQ(numFramesPlayed & (1 << 31), 0);
int64_t durationUs =
getDurationIfPlayedAtNativeSampleRate_l(numFramesPlayed)
+ nowUs - numFramesPlayedAtUs;
if (durationUs < 0) {
// Occurs when numFramesPlayed position is very small and the following:
// (1) In case 1, the time nowUs is computed before getTimestamp() is
// called and numFramesPlayedAtUs is greater than nowUs by time more
// than numFramesPlayed.
// (2) In case 3, using getPosition and adding mAudioTrack->latency()
// to numFramesPlayedAtUs, by a time amount greater than
// numFramesPlayed.
//
// Both of these are transitory conditions.
ALOGV("getPlayedOutAudioDurationMedia_l: negative duration %lld "
"set to zero", (long long)durationUs);
durationUs = 0;
}
ALOGV("getPlayedOutAudioDurationMedia_l(%lld) nowUs(%lld) frames(%u) "
"framesAt(%lld)",
(long long)durationUs, (long long)nowUs, numFramesPlayed,
(long long)numFramesPlayedAtUs);
return durationUs;
}
void MediaSync::onDrainVideo_l() {
if (!isPlaying()) {
return;
}
while (!mBufferItems.empty()) {
int64_t nowUs = ALooper::GetNowUs();
BufferItem *bufferItem = &*mBufferItems.begin();
int64_t itemMediaUs = bufferItem->mTimestamp / 1000;
int64_t itemRealUs = getRealTime(itemMediaUs, nowUs);
// adjust video frame PTS based on vsync
itemRealUs = mFrameScheduler->schedule(itemRealUs * 1000) / 1000;
int64_t twoVsyncsUs = 2 * (mFrameScheduler->getVsyncPeriod() / 1000);
// post 2 display refreshes before rendering is due
if (itemRealUs <= nowUs + twoVsyncsUs) {
ALOGV("adjusting PTS from %lld to %lld",
(long long)bufferItem->mTimestamp / 1000, (long long)itemRealUs);
bufferItem->mTimestamp = itemRealUs * 1000;
bufferItem->mIsAutoTimestamp = false;
if (mHasAudio) {
if (nowUs - itemRealUs <= kMaxAllowedVideoLateTimeUs) {
renderOneBufferItem_l(*bufferItem);
} else {
// too late.
returnBufferToInput_l(
bufferItem->mGraphicBuffer, bufferItem->mFence);
mFrameScheduler->restart();
}
} else {
// always render video buffer in video-only mode.
renderOneBufferItem_l(*bufferItem);
// smooth out videos >= 10fps
mMediaClock->updateAnchor(
itemMediaUs, nowUs, itemMediaUs + 100000);
}
mBufferItems.erase(mBufferItems.begin());
mNextBufferItemMediaUs = -1;
} else {
if (mNextBufferItemMediaUs == -1
|| mNextBufferItemMediaUs > itemMediaUs) {
sp<AMessage> msg = new AMessage(kWhatDrainVideo, this);
msg->post(itemRealUs - nowUs - twoVsyncsUs);
mNextBufferItemMediaUs = itemMediaUs;
}
break;
}
}
}
void MediaSync::onFrameAvailableFromInput() {
Mutex::Autolock lock(mMutex);
const static nsecs_t kAcquireWaitTimeout = 2000000000; // 2 seconds
mReturnPendingInputFrame = false;
// If there are too many outstanding buffers, wait until a buffer is
// released back to the input in onBufferReleased.
// NOTE: BufferQueue allows dequeuing maxAcquiredBufferCount + 1 buffers
while (mNumOutstandingBuffers > mMaxAcquiredBufferCount
&& !mIsAbandoned && !mReturnPendingInputFrame) {
if (mReleaseCondition.waitRelative(mMutex, kAcquireWaitTimeout) != OK) {
ALOGI_IF(mPlaybackRate != 0.f, "still waiting to release a buffer before acquire");
}
// If the sync is abandoned while we are waiting, the release
// condition variable will be broadcast, and we should just return
// without attempting to do anything more (since the input queue will
// also be abandoned).
if (mIsAbandoned) {
return;
}
}
// Acquire and detach the buffer from the input.
BufferItem bufferItem;
status_t status = mInput->acquireBuffer(&bufferItem, 0 /* presentWhen */);
if (status != NO_ERROR) {
ALOGE("acquiring buffer from input failed (%d)", status);
return;
}
++mNumOutstandingBuffers;
ALOGV("acquired buffer %#llx from input", (long long)bufferItem.mGraphicBuffer->getId());
status = mInput->detachBuffer(bufferItem.mSlot);
if (status != NO_ERROR) {
ALOGE("detaching buffer from input failed (%d)", status);
if (status == NO_INIT) {
// If the input has been abandoned, move on.
onAbandoned_l(true /* isInput */);
}
return;
}
if (mBuffersFromInput.indexOfKey(bufferItem.mGraphicBuffer->getId()) >= 0) {
// Something is wrong since this buffer should be at our hands, bail.
ALOGE("received buffer multiple times from input");
mInput->consumerDisconnect();
onAbandoned_l(true /* isInput */);
return;
}
mBuffersFromInput.add(bufferItem.mGraphicBuffer->getId(), bufferItem.mGraphicBuffer);
// If flush happened while waiting for a buffer to be released, simply return it
// TRICKY: do it here after it is detached so that we don't have to cache mGraphicBuffer.
if (mReturnPendingInputFrame) {
mReturnPendingInputFrame = false;
returnBufferToInput_l(bufferItem.mGraphicBuffer, bufferItem.mFence);
return;
}
mBufferItems.push_back(bufferItem);
if (mBufferItems.size() == 1) {
onDrainVideo_l();
}
}
void MediaSync::renderOneBufferItem_l(const BufferItem &bufferItem) {
IGraphicBufferProducer::QueueBufferInput queueInput(
bufferItem.mTimestamp,
bufferItem.mIsAutoTimestamp,
bufferItem.mDataSpace,
bufferItem.mCrop,
static_cast<int32_t>(bufferItem.mScalingMode),
bufferItem.mTransform,
bufferItem.mFence);
// Attach and queue the buffer to the output.
int slot;
mOutput->setGenerationNumber(bufferItem.mGraphicBuffer->getGenerationNumber());
status_t status = mOutput->attachBuffer(&slot, bufferItem.mGraphicBuffer);
ALOGE_IF(status != NO_ERROR, "attaching buffer to output failed (%d)", status);
if (status == NO_ERROR) {
IGraphicBufferProducer::QueueBufferOutput queueOutput;
status = mOutput->queueBuffer(slot, queueInput, &queueOutput);
ALOGE_IF(status != NO_ERROR, "queueing buffer to output failed (%d)", status);
}
if (status != NO_ERROR) {
returnBufferToInput_l(bufferItem.mGraphicBuffer, bufferItem.mFence);
if (status == NO_INIT) {
// If the output has been abandoned, move on.
onAbandoned_l(false /* isInput */);
}
return;
}
if (mBuffersSentToOutput.indexOfKey(bufferItem.mGraphicBuffer->getId()) >= 0) {
// Something is wrong since this buffer should be held by output now, bail.
mInput->consumerDisconnect();
onAbandoned_l(true /* isInput */);
return;
}
mBuffersSentToOutput.add(bufferItem.mGraphicBuffer->getId(), bufferItem.mGraphicBuffer);
ALOGV("queued buffer %#llx to output", (long long)bufferItem.mGraphicBuffer->getId());
}
void MediaSync::onBufferReleasedByOutput(sp<IGraphicBufferProducer> &output) {
Mutex::Autolock lock(mMutex);
if (output != mOutput) {
return; // This is not the current output, ignore.
}
sp<GraphicBuffer> buffer;
sp<Fence> fence;
status_t status = mOutput->detachNextBuffer(&buffer, &fence);
ALOGE_IF(status != NO_ERROR, "detaching buffer from output failed (%d)", status);
if (status == NO_INIT) {
// If the output has been abandoned, we can't do anything else,
// since buffer is invalid.
onAbandoned_l(false /* isInput */);
return;
}
ALOGV("detached buffer %#llx from output", (long long)buffer->getId());
// If we've been abandoned, we can't return the buffer to the input, so just
// move on.
if (mIsAbandoned) {
return;
}
ssize_t ix = mBuffersSentToOutput.indexOfKey(buffer->getId());
if (ix < 0) {
// The buffer is unknown, maybe leftover, ignore.
return;
}
mBuffersSentToOutput.removeItemsAt(ix);
returnBufferToInput_l(buffer, fence);
}
void MediaSync::returnBufferToInput_l(
const sp<GraphicBuffer> &buffer, const sp<Fence> &fence) {
ssize_t ix = mBuffersFromInput.indexOfKey(buffer->getId());
if (ix < 0) {
// The buffer is unknown, something is wrong, bail.
ALOGE("output returned unknown buffer");
mOutput->disconnect(NATIVE_WINDOW_API_MEDIA);
onAbandoned_l(false /* isInput */);
return;
}
sp<GraphicBuffer> oldBuffer = mBuffersFromInput.valueAt(ix);
mBuffersFromInput.removeItemsAt(ix);
// Attach and release the buffer back to the input.
int consumerSlot;
status_t status = mInput->attachBuffer(&consumerSlot, oldBuffer);
ALOGE_IF(status != NO_ERROR, "attaching buffer to input failed (%d)", status);
if (status == NO_ERROR) {
status = mInput->releaseBuffer(consumerSlot, 0 /* frameNumber */,
EGL_NO_DISPLAY, EGL_NO_SYNC_KHR, fence);
ALOGE_IF(status != NO_ERROR, "releasing buffer to input failed (%d)", status);
}
// Notify any waiting onFrameAvailable calls.
--mNumOutstandingBuffers;
mReleaseCondition.signal();
if (status == NO_ERROR) {
ALOGV("released buffer %#llx to input", (long long)oldBuffer->getId());
}
}
void MediaSync::onAbandoned_l(bool isInput) {
ALOGE("the %s has abandoned me", (isInput ? "input" : "output"));
if (!mIsAbandoned) {
if (isInput) {
mOutput->disconnect(NATIVE_WINDOW_API_MEDIA);
} else {
mInput->consumerDisconnect();
}
mIsAbandoned = true;
}
mReleaseCondition.broadcast();
}
void MediaSync::onMessageReceived(const sp<AMessage> &msg) {
switch (msg->what()) {
case kWhatDrainVideo:
{
Mutex::Autolock lock(mMutex);
if (mNextBufferItemMediaUs != -1) {
int64_t nowUs = ALooper::GetNowUs();
int64_t itemRealUs = getRealTime(mNextBufferItemMediaUs, nowUs);
// The message could arrive earlier than expected due to
// various reasons, e.g., media clock has been changed because
// of new anchor time or playback rate. In such cases, the
// message needs to be re-posted.
if (itemRealUs > nowUs) {
msg->post(itemRealUs - nowUs);
break;
}
}
onDrainVideo_l();
break;
}
default:
TRESPASS();
break;
}
}
MediaSync::InputListener::InputListener(const sp<MediaSync> &sync)
: mSync(sync) {}
MediaSync::InputListener::~InputListener() {}
void MediaSync::InputListener::onFrameAvailable(const BufferItem &/* item */) {
mSync->onFrameAvailableFromInput();
}
// We don't care about sideband streams, since we won't relay them.
void MediaSync::InputListener::onSidebandStreamChanged() {
ALOGE("onSidebandStreamChanged: got sideband stream unexpectedly.");
}
void MediaSync::InputListener::binderDied(const wp<IBinder> &/* who */) {
Mutex::Autolock lock(mSync->mMutex);
mSync->onAbandoned_l(true /* isInput */);
}
MediaSync::OutputListener::OutputListener(const sp<MediaSync> &sync,
const sp<IGraphicBufferProducer> &output)
: mSync(sync),
mOutput(output) {}
MediaSync::OutputListener::~OutputListener() {}
void MediaSync::OutputListener::onBufferReleased() {
mSync->onBufferReleasedByOutput(mOutput);
}
void MediaSync::OutputListener::binderDied(const wp<IBinder> &/* who */) {
Mutex::Autolock lock(mSync->mMutex);
mSync->onAbandoned_l(false /* isInput */);
}
} // namespace android