blob: 1a50ccb5707343aca1a65bf53181b1e793975e7b [file] [log] [blame]
/*
* Copyright (C) 2011 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
//#define USE_LOG SLAndroidLogLevel_Verbose
#include "sles_allinclusive.h"
#include "android/android_AudioSfDecoder.h"
#include <binder/IServiceManager.h>
#include <media/stagefright/foundation/ADebug.h>
#define SIZE_CACHED_HIGH_BYTES 1000000
#define SIZE_CACHED_MED_BYTES 700000
#define SIZE_CACHED_LOW_BYTES 400000
namespace android {
//--------------------------------------------------------------------------------------------------
AudioSfDecoder::AudioSfDecoder(const AudioPlayback_Parameters* params) : GenericPlayer(params),
mDataSource(0),
mAudioSource(0),
mAudioSourceStarted(false),
mBitrate(-1),
mDurationUsec(ANDROID_UNKNOWN_TIME),
mDecodeBuffer(NULL),
mSeekTimeMsec(0),
// play event logic depends on the initial time being zero not ANDROID_UNKNOWN_TIME
mLastDecodedPositionUs(0)
{
SL_LOGD("AudioSfDecoder::AudioSfDecoder()");
}
AudioSfDecoder::~AudioSfDecoder() {
SL_LOGD("AudioSfDecoder::~AudioSfDecoder()");
}
void AudioSfDecoder::preDestroy() {
GenericPlayer::preDestroy();
SL_LOGD("AudioSfDecoder::preDestroy()");
{
Mutex::Autolock _l(mBufferSourceLock);
if (NULL != mDecodeBuffer) {
mDecodeBuffer->release();
mDecodeBuffer = NULL;
}
if ((mAudioSource != 0) && mAudioSourceStarted) {
mAudioSource->stop();
mAudioSourceStarted = false;
}
}
}
//--------------------------------------------------
void AudioSfDecoder::play() {
SL_LOGD("AudioSfDecoder::play");
GenericPlayer::play();
(new AMessage(kWhatDecode, id()))->post();
}
void AudioSfDecoder::getPositionMsec(int* msec) {
int64_t timeUsec = getPositionUsec();
if (timeUsec == ANDROID_UNKNOWN_TIME) {
*msec = ANDROID_UNKNOWN_TIME;
} else {
*msec = timeUsec / 1000;
}
}
//--------------------------------------------------
uint32_t AudioSfDecoder::getPcmFormatKeyCount() const {
return NB_PCMMETADATA_KEYS;
}
//--------------------------------------------------
bool AudioSfDecoder::getPcmFormatKeySize(uint32_t index, uint32_t* pKeySize) {
if (index >= NB_PCMMETADATA_KEYS) {
return false;
} else {
*pKeySize = strlen(kPcmDecodeMetadataKeys[index]) +1;
return true;
}
}
//--------------------------------------------------
bool AudioSfDecoder::getPcmFormatKeyName(uint32_t index, uint32_t keySize, char* keyName) {
uint32_t actualKeySize;
if (!getPcmFormatKeySize(index, &actualKeySize)) {
return false;
}
if (keySize < actualKeySize) {
return false;
}
strncpy(keyName, kPcmDecodeMetadataKeys[index], actualKeySize);
return true;
}
//--------------------------------------------------
bool AudioSfDecoder::getPcmFormatValueSize(uint32_t index, uint32_t* pValueSize) {
if (index >= NB_PCMMETADATA_KEYS) {
*pValueSize = 0;
return false;
} else {
*pValueSize = sizeof(uint32_t);
return true;
}
}
//--------------------------------------------------
bool AudioSfDecoder::getPcmFormatKeyValue(uint32_t index, uint32_t size, uint32_t* pValue) {
uint32_t valueSize = 0;
if (!getPcmFormatValueSize(index, &valueSize)) {
return false;
} else if (size != valueSize) {
// this ensures we are accessing mPcmFormatValues with a valid size for that index
SL_LOGE("Error retrieving metadata value at index %d: using size of %d, should be %d",
index, size, valueSize);
return false;
} else {
android::Mutex::Autolock autoLock(mPcmFormatLock);
*pValue = mPcmFormatValues[index];
return true;
}
}
//--------------------------------------------------
// Event handlers
// it is strictly verboten to call those methods outside of the event loop
// Initializes the data and audio sources, and update the PCM format info
// post-condition: upon successful initialization based on the player data locator
// GenericPlayer::onPrepare() was called
// mDataSource != 0
// mAudioSource != 0
// mAudioSourceStarted == true
// All error returns from this method are via notifyPrepared(status) followed by "return".
void AudioSfDecoder::onPrepare() {
SL_LOGD("AudioSfDecoder::onPrepare()");
Mutex::Autolock _l(mBufferSourceLock);
{
android::Mutex::Autolock autoLock(mPcmFormatLock);
// Initialize the PCM format info with the known parameters before the start of the decode
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_BITSPERSAMPLE] = SL_PCMSAMPLEFORMAT_FIXED_16;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_CONTAINERSIZE] = 16;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_ENDIANNESS] = SL_BYTEORDER_LITTLEENDIAN;
// initialization with the default values: they will be replaced by the actual values
// once the decoder has figured them out
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_NUMCHANNELS] = UNKNOWN_NUMCHANNELS;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_SAMPLERATE] = UNKNOWN_SAMPLERATE;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_CHANNELMASK] = UNKNOWN_CHANNELMASK;
}
//---------------------------------
// Instantiate and initialize the data source for the decoder
sp<DataSource> dataSource;
switch (mDataLocatorType) {
case kDataLocatorNone:
SL_LOGE("AudioSfDecoder::onPrepare: no data locator set");
notifyPrepared(MEDIA_ERROR_BASE);
return;
case kDataLocatorUri:
dataSource = DataSource::CreateFromURI(mDataLocator.uriRef);
if (dataSource == NULL) {
SL_LOGE("AudioSfDecoder::onPrepare(): Error opening %s", mDataLocator.uriRef);
notifyPrepared(MEDIA_ERROR_BASE);
return;
}
break;
case kDataLocatorFd:
{
// As FileSource unconditionally takes ownership of the fd and closes it, then
// we have to make a dup for FileSource if the app wants to keep ownership itself
int fd = mDataLocator.fdi.fd;
if (mDataLocator.fdi.mCloseAfterUse) {
mDataLocator.fdi.mCloseAfterUse = false;
} else {
fd = ::dup(fd);
}
dataSource = new FileSource(fd, mDataLocator.fdi.offset, mDataLocator.fdi.length);
status_t err = dataSource->initCheck();
if (err != OK) {
notifyPrepared(err);
return;
}
break;
}
// AndroidBufferQueue data source is handled by a subclass,
// which does not call up to this method. Hence, the missing case.
default:
TRESPASS();
}
//---------------------------------
// Instantiate and initialize the decoder attached to the data source
sp<MediaExtractor> extractor = MediaExtractor::Create(dataSource);
if (extractor == NULL) {
SL_LOGE("AudioSfDecoder::onPrepare: Could not instantiate extractor.");
notifyPrepared(ERROR_UNSUPPORTED);
return;
}
ssize_t audioTrackIndex = -1;
bool isRawAudio = false;
for (size_t i = 0; i < extractor->countTracks(); ++i) {
sp<MetaData> meta = extractor->getTrackMetaData(i);
const char *mime;
CHECK(meta->findCString(kKeyMIMEType, &mime));
if (!strncasecmp("audio/", mime, 6)) {
if (isSupportedCodec(mime)) {
audioTrackIndex = i;
if (!strcasecmp(MEDIA_MIMETYPE_AUDIO_RAW, mime)) {
isRawAudio = true;
}
break;
}
}
}
if (audioTrackIndex < 0) {
SL_LOGE("AudioSfDecoder::onPrepare: Could not find a supported audio track.");
notifyPrepared(ERROR_UNSUPPORTED);
return;
}
sp<MediaSource> source = extractor->getTrack(audioTrackIndex);
sp<MetaData> meta = source->getFormat();
// we can't trust the OMXCodec (if there is one) to issue a INFO_FORMAT_CHANGED so we want
// to have some meaningful values as soon as possible.
int32_t channelCount;
bool hasChannelCount = meta->findInt32(kKeyChannelCount, &channelCount);
int32_t sr;
bool hasSampleRate = meta->findInt32(kKeySampleRate, &sr);
// first compute the duration
off64_t size;
int64_t durationUs;
int32_t durationMsec;
if (dataSource->getSize(&size) == OK
&& meta->findInt64(kKeyDuration, &durationUs)) {
if (durationUs != 0) {
mBitrate = size * 8000000ll / durationUs; // in bits/sec
} else {
mBitrate = -1;
}
mDurationUsec = durationUs;
durationMsec = durationUs / 1000;
} else {
mBitrate = -1;
mDurationUsec = ANDROID_UNKNOWN_TIME;
durationMsec = ANDROID_UNKNOWN_TIME;
}
// then assign the duration under the settings lock
{
Mutex::Autolock _l(mSettingsLock);
mDurationMsec = durationMsec;
}
// the audio content is not raw PCM, so we need a decoder
if (!isRawAudio) {
OMXClient client;
CHECK_EQ(client.connect(), (status_t)OK);
source = OMXCodec::Create(
client.interface(), meta, false /* createEncoder */,
source);
if (source == NULL) {
SL_LOGE("AudioSfDecoder::onPrepare: Could not instantiate decoder.");
notifyPrepared(ERROR_UNSUPPORTED);
return;
}
meta = source->getFormat();
}
if (source->start() != OK) {
SL_LOGE("AudioSfDecoder::onPrepare: Failed to start source/decoder.");
notifyPrepared(MEDIA_ERROR_BASE);
return;
}
//---------------------------------
// The data source, and audio source (a decoder if required) are ready to be used
mDataSource = dataSource;
mAudioSource = source;
mAudioSourceStarted = true;
if (!hasChannelCount) {
CHECK(meta->findInt32(kKeyChannelCount, &channelCount));
}
if (!hasSampleRate) {
CHECK(meta->findInt32(kKeySampleRate, &sr));
}
// FIXME add code below once channel mask support is in, currently initialized to default
// value computed from the channel count
// if (!hasChannelMask) {
// CHECK(meta->findInt32(kKeyChannelMask, &channelMask));
// }
if (!wantPrefetch()) {
SL_LOGV("AudioSfDecoder::onPrepare: no need to prefetch");
// doesn't need prefetching, notify good to go
mCacheStatus = kStatusHigh;
mCacheFill = 1000;
notifyStatus();
notifyCacheFill();
}
{
android::Mutex::Autolock autoLock(mPcmFormatLock);
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_SAMPLERATE] = sr;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_NUMCHANNELS] = channelCount;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_CHANNELMASK] =
channelCountToMask(channelCount);
}
// at this point we have enough information about the source to create the sink that
// will consume the data
createAudioSink();
// signal successful completion of prepare
mStateFlags |= kFlagPrepared;
GenericPlayer::onPrepare();
SL_LOGD("AudioSfDecoder::onPrepare() done, mStateFlags=0x%x", mStateFlags);
}
void AudioSfDecoder::onPause() {
SL_LOGV("AudioSfDecoder::onPause()");
GenericPlayer::onPause();
pauseAudioSink();
}
void AudioSfDecoder::onPlay() {
SL_LOGV("AudioSfDecoder::onPlay()");
GenericPlayer::onPlay();
startAudioSink();
}
void AudioSfDecoder::onSeek(const sp<AMessage> &msg) {
SL_LOGV("AudioSfDecoder::onSeek");
int64_t timeMsec;
CHECK(msg->findInt64(WHATPARAM_SEEK_SEEKTIME_MS, &timeMsec));
Mutex::Autolock _l(mTimeLock);
mStateFlags |= kFlagSeeking;
mSeekTimeMsec = timeMsec;
// don't set mLastDecodedPositionUs to ANDROID_UNKNOWN_TIME; getPositionUsec
// ignores mLastDecodedPositionUs while seeking, and substitutes the seek goal instead
// nop for now
GenericPlayer::onSeek(msg);
}
void AudioSfDecoder::onLoop(const sp<AMessage> &msg) {
SL_LOGV("AudioSfDecoder::onLoop");
int32_t loop;
CHECK(msg->findInt32(WHATPARAM_LOOP_LOOPING, &loop));
if (loop) {
//SL_LOGV("AudioSfDecoder::onLoop start looping");
mStateFlags |= kFlagLooping;
} else {
//SL_LOGV("AudioSfDecoder::onLoop stop looping");
mStateFlags &= ~kFlagLooping;
}
// nop for now
GenericPlayer::onLoop(msg);
}
void AudioSfDecoder::onCheckCache(const sp<AMessage> &msg) {
//SL_LOGV("AudioSfDecoder::onCheckCache");
bool eos;
CacheStatus_t status = getCacheRemaining(&eos);
if (eos || status == kStatusHigh
|| ((mStateFlags & kFlagPreparing) && (status >= kStatusEnough))) {
if (mStateFlags & kFlagPlaying) {
startAudioSink();
}
mStateFlags &= ~kFlagBuffering;
SL_LOGV("AudioSfDecoder::onCheckCache: buffering done.");
if (mStateFlags & kFlagPreparing) {
//SL_LOGV("AudioSfDecoder::onCheckCache: preparation done.");
mStateFlags &= ~kFlagPreparing;
}
if (mStateFlags & kFlagPlaying) {
(new AMessage(kWhatDecode, id()))->post();
}
return;
}
msg->post(100000);
}
void AudioSfDecoder::onDecode() {
SL_LOGV("AudioSfDecoder::onDecode");
//-------------------------------- Need to buffer some more before decoding?
bool eos;
if (mDataSource == 0) {
// application set play state to paused which failed, then set play state to playing
return;
}
if (wantPrefetch()
&& (getCacheRemaining(&eos) == kStatusLow)
&& !eos) {
SL_LOGV("buffering more.");
if (mStateFlags & kFlagPlaying) {
pauseAudioSink();
}
mStateFlags |= kFlagBuffering;
(new AMessage(kWhatCheckCache, id()))->post(100000);
return;
}
if (!(mStateFlags & (kFlagPlaying | kFlagBuffering | kFlagPreparing))) {
// don't decode if we're not buffering, prefetching or playing
//SL_LOGV("don't decode: not buffering, prefetching or playing");
return;
}
//-------------------------------- Decode
status_t err;
MediaSource::ReadOptions readOptions;
if (mStateFlags & kFlagSeeking) {
assert(mSeekTimeMsec != ANDROID_UNKNOWN_TIME);
readOptions.setSeekTo(mSeekTimeMsec * 1000);
}
int64_t timeUsec = ANDROID_UNKNOWN_TIME;
{
Mutex::Autolock _l(mBufferSourceLock);
if (NULL != mDecodeBuffer) {
// the current decoded buffer hasn't been rendered, drop it
mDecodeBuffer->release();
mDecodeBuffer = NULL;
}
if (!mAudioSourceStarted) {
return;
}
err = mAudioSource->read(&mDecodeBuffer, &readOptions);
if (err == OK) {
// FIXME workaround apparent bug in AAC decoder: kKeyTime is 3 frames old if length is 0
if (mDecodeBuffer->range_length() == 0) {
timeUsec = ANDROID_UNKNOWN_TIME;
} else {
CHECK(mDecodeBuffer->meta_data()->findInt64(kKeyTime, &timeUsec));
}
} else {
// errors are handled below
}
}
{
Mutex::Autolock _l(mTimeLock);
if (mStateFlags & kFlagSeeking) {
mStateFlags &= ~kFlagSeeking;
mSeekTimeMsec = ANDROID_UNKNOWN_TIME;
}
if (timeUsec != ANDROID_UNKNOWN_TIME) {
// Note that though we've decoded this position, we haven't rendered it yet.
// So a GetPosition called after this point will observe the advanced position,
// even though the PCM may not have been supplied to the sink. That's OK as
// we don't claim to provide AAC frame-accurate (let alone sample-accurate) GetPosition.
mLastDecodedPositionUs = timeUsec;
}
}
//-------------------------------- Handle return of decode
if (err != OK) {
bool continueDecoding = false;
switch(err) {
case ERROR_END_OF_STREAM:
if (0 < mDurationUsec) {
Mutex::Autolock _l(mTimeLock);
mLastDecodedPositionUs = mDurationUsec;
}
// handle notification and looping at end of stream
if (mStateFlags & kFlagPlaying) {
notify(PLAYEREVENT_ENDOFSTREAM, 1, true /*async*/);
}
if (mStateFlags & kFlagLooping) {
seek(0);
// kick-off decoding again
continueDecoding = true;
}
break;
case INFO_FORMAT_CHANGED:
SL_LOGD("MediaSource::read encountered INFO_FORMAT_CHANGED");
// reconfigure output
{
Mutex::Autolock _l(mBufferSourceLock);
hasNewDecodeParams();
}
continueDecoding = true;
break;
case INFO_DISCONTINUITY:
SL_LOGD("MediaSource::read encountered INFO_DISCONTINUITY");
continueDecoding = true;
break;
default:
SL_LOGE("MediaSource::read returned error %d", err);
break;
}
if (continueDecoding) {
if (NULL == mDecodeBuffer) {
(new AMessage(kWhatDecode, id()))->post();
return;
}
} else {
return;
}
}
//-------------------------------- Render
sp<AMessage> msg = new AMessage(kWhatRender, id());
msg->post();
}
void AudioSfDecoder::onMessageReceived(const sp<AMessage> &msg) {
switch (msg->what()) {
case kWhatDecode:
onDecode();
break;
case kWhatRender:
onRender();
break;
case kWhatCheckCache:
onCheckCache(msg);
break;
default:
GenericPlayer::onMessageReceived(msg);
break;
}
}
//--------------------------------------------------
// Prepared state, prefetch status notifications
void AudioSfDecoder::notifyPrepared(status_t prepareRes) {
assert(!(mStateFlags & (kFlagPrepared | kFlagPreparedUnsuccessfully)));
if (NO_ERROR == prepareRes) {
// The "then" fork is not currently used, but is kept here to make it easier
// to replace by a new signalPrepareCompletion(status) if we re-visit this later.
mStateFlags |= kFlagPrepared;
} else {
mStateFlags |= kFlagPreparedUnsuccessfully;
}
// Do not call the superclass onPrepare to notify, because it uses a default error
// status code but we can provide a more specific one.
// GenericPlayer::onPrepare();
notify(PLAYEREVENT_PREPARED, (int32_t)prepareRes, true /*async*/);
SL_LOGD("AudioSfDecoder::onPrepare() done, mStateFlags=0x%x", mStateFlags);
}
void AudioSfDecoder::onNotify(const sp<AMessage> &msg) {
notif_cbf_t notifyClient;
void* notifyUser;
{
android::Mutex::Autolock autoLock(mNotifyClientLock);
if (NULL == mNotifyClient) {
return;
} else {
notifyClient = mNotifyClient;
notifyUser = mNotifyUser;
}
}
int32_t val;
if (msg->findInt32(PLAYEREVENT_PREFETCHSTATUSCHANGE, &val)) {
SL_LOGV("\tASfPlayer notifying %s = %d", PLAYEREVENT_PREFETCHSTATUSCHANGE, val);
notifyClient(kEventPrefetchStatusChange, val, 0, notifyUser);
}
else if (msg->findInt32(PLAYEREVENT_PREFETCHFILLLEVELUPDATE, &val)) {
SL_LOGV("\tASfPlayer notifying %s = %d", PLAYEREVENT_PREFETCHFILLLEVELUPDATE, val);
notifyClient(kEventPrefetchFillLevelUpdate, val, 0, notifyUser);
}
else if (msg->findInt32(PLAYEREVENT_ENDOFSTREAM, &val)) {
SL_LOGV("\tASfPlayer notifying %s = %d", PLAYEREVENT_ENDOFSTREAM, val);
notifyClient(kEventEndOfStream, val, 0, notifyUser);
}
else {
GenericPlayer::onNotify(msg);
}
}
//--------------------------------------------------
// Private utility functions
bool AudioSfDecoder::wantPrefetch() {
if (mDataSource != 0) {
return (mDataSource->flags() & DataSource::kWantsPrefetching);
} else {
// happens if an improper data locator was passed, if the media extractor couldn't be
// initialized, if there is no audio track in the media, if the OMX decoder couldn't be
// instantiated, if the source couldn't be opened, or if the MediaSource
// couldn't be started
SL_LOGV("AudioSfDecoder::wantPrefetch() tries to access NULL mDataSource");
return false;
}
}
int64_t AudioSfDecoder::getPositionUsec() {
Mutex::Autolock _l(mTimeLock);
if (mStateFlags & kFlagSeeking) {
return mSeekTimeMsec * 1000;
} else {
return mLastDecodedPositionUs;
}
}
CacheStatus_t AudioSfDecoder::getCacheRemaining(bool *eos) {
sp<NuCachedSource2> cachedSource =
static_cast<NuCachedSource2 *>(mDataSource.get());
CacheStatus_t oldStatus = mCacheStatus;
status_t finalStatus;
size_t dataRemaining = cachedSource->approxDataRemaining(&finalStatus);
*eos = (finalStatus != OK);
CHECK_GE(mBitrate, 0);
int64_t dataRemainingUs = dataRemaining * 8000000ll / mBitrate;
//SL_LOGV("AudioSfDecoder::getCacheRemaining: approx %.2f secs remaining (eos=%d)",
// dataRemainingUs / 1E6, *eos);
if (*eos) {
// data is buffered up to the end of the stream, it can't get any better than this
mCacheStatus = kStatusHigh;
mCacheFill = 1000;
} else {
if (mDurationUsec > 0) {
// known duration:
// fill level is ratio of how much has been played + how much is
// cached, divided by total duration
int64_t currentPositionUsec = getPositionUsec();
if (currentPositionUsec == ANDROID_UNKNOWN_TIME) {
// if we don't know where we are, assume the worst for the fill ratio
currentPositionUsec = 0;
}
if (mDurationUsec > 0) {
mCacheFill = (int16_t) ((1000.0
* (double)(currentPositionUsec + dataRemainingUs) / mDurationUsec));
} else {
mCacheFill = 0;
}
//SL_LOGV("cacheFill = %d", mCacheFill);
// cache status is evaluated against duration thresholds
if (dataRemainingUs > DURATION_CACHED_HIGH_MS*1000) {
mCacheStatus = kStatusHigh;
//ALOGV("high");
} else if (dataRemainingUs > DURATION_CACHED_MED_MS*1000) {
//ALOGV("enough");
mCacheStatus = kStatusEnough;
} else if (dataRemainingUs < DURATION_CACHED_LOW_MS*1000) {
//ALOGV("low");
mCacheStatus = kStatusLow;
} else {
mCacheStatus = kStatusIntermediate;
}
} else {
// unknown duration:
// cache status is evaluated against cache amount thresholds
// (no duration so we don't have the bitrate either, could be derived from format?)
if (dataRemaining > SIZE_CACHED_HIGH_BYTES) {
mCacheStatus = kStatusHigh;
} else if (dataRemaining > SIZE_CACHED_MED_BYTES) {
mCacheStatus = kStatusEnough;
} else if (dataRemaining < SIZE_CACHED_LOW_BYTES) {
mCacheStatus = kStatusLow;
} else {
mCacheStatus = kStatusIntermediate;
}
}
}
if (oldStatus != mCacheStatus) {
notifyStatus();
}
if (abs(mCacheFill - mLastNotifiedCacheFill) > mCacheFillNotifThreshold) {
notifyCacheFill();
}
return mCacheStatus;
}
void AudioSfDecoder::hasNewDecodeParams() {
if ((mAudioSource != 0) && mAudioSourceStarted) {
sp<MetaData> meta = mAudioSource->getFormat();
int32_t channelCount;
CHECK(meta->findInt32(kKeyChannelCount, &channelCount));
int32_t sr;
CHECK(meta->findInt32(kKeySampleRate, &sr));
// FIXME similar to onPrepare()
{
android::Mutex::Autolock autoLock(mPcmFormatLock);
SL_LOGV("format changed: old sr=%d, channels=%d; new sr=%d, channels=%d",
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_SAMPLERATE],
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_NUMCHANNELS],
sr, channelCount);
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_NUMCHANNELS] = channelCount;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_SAMPLERATE] = sr;
mPcmFormatValues[ANDROID_KEY_INDEX_PCMFORMAT_CHANNELMASK] =
channelCountToMask(channelCount);
}
// there's no need to do a notify of PLAYEREVENT_CHANNEL_COUNT,
// because the only listener is for volume updates, and decoders don't support that
}
// alert users of those params
updateAudioSink();
}
static const char* const kPlaybackOnlyCodecs[] = { MEDIA_MIMETYPE_AUDIO_AMR_NB,
MEDIA_MIMETYPE_AUDIO_AMR_WB };
#define NB_PLAYBACK_ONLY_CODECS (sizeof(kPlaybackOnlyCodecs)/sizeof(kPlaybackOnlyCodecs[0]))
bool AudioSfDecoder::isSupportedCodec(const char* mime) {
bool codecRequiresPermission = false;
for (unsigned int i = 0 ; i < NB_PLAYBACK_ONLY_CODECS ; i++) {
if (!strcasecmp(mime, kPlaybackOnlyCodecs[i])) {
codecRequiresPermission = true;
break;
}
}
if (codecRequiresPermission) {
// verify only the system can decode, for playback only
return checkCallingPermission(
String16("android.permission.ALLOW_ANY_CODEC_FOR_PLAYBACK"));
} else {
return true;
}
}
} // namespace android