blob: cadbb11bc155825dc0901da6c7d59c241ed56210 [file] [log] [blame]
/*
* Copyright (c) 2002, 2015, Oracle and/or its affiliates. All rights reserved.
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
*
* This code is free software; you can redistribute it and/or modify it
* under the terms of the GNU General Public License version 2 only, as
* published by the Free Software Foundation. Oracle designates this
* particular file as subject to the "Classpath" exception as provided
* by Oracle in the LICENSE file that accompanied this code.
*
* This code is distributed in the hope that it will be useful, but WITHOUT
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License
* version 2 for more details (a copy is included in the LICENSE file that
* accompanied this code).
*
* You should have received a copy of the GNU General Public License version
* 2 along with this work; if not, write to the Free Software Foundation,
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
*
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
* or visit www.oracle.com if you need additional information or have any
* questions.
*/
//#define USE_ERROR
//#define USE_TRACE
//#define USE_VERBOSE_TRACE
#include <AudioUnit/AudioUnit.h>
#include <AudioToolbox/AudioConverter.h>
#include <pthread.h>
#include <math.h>
/*
#if !defined(__COREAUDIO_USE_FLAT_INCLUDES__)
#include <CoreAudio/CoreAudioTypes.h>
#else
#include <CoreAudioTypes.h>
#endif
*/
#include "PLATFORM_API_MacOSX_Utils.h"
extern "C" {
#include "Utilities.h"
#include "DirectAudio.h"
}
#if USE_DAUDIO == TRUE
#ifdef USE_TRACE
static void PrintStreamDesc(const AudioStreamBasicDescription *inDesc) {
TRACE4("ID='%c%c%c%c'", (char)(inDesc->mFormatID >> 24), (char)(inDesc->mFormatID >> 16), (char)(inDesc->mFormatID >> 8), (char)(inDesc->mFormatID));
TRACE2(", %f Hz, flags=0x%lX", (float)inDesc->mSampleRate, (long unsigned)inDesc->mFormatFlags);
TRACE2(", %ld channels, %ld bits", (long)inDesc->mChannelsPerFrame, (long)inDesc->mBitsPerChannel);
TRACE1(", %ld bytes per frame\n", (long)inDesc->mBytesPerFrame);
}
#else
static inline void PrintStreamDesc(const AudioStreamBasicDescription *inDesc) { }
#endif
#define MAX(x, y) ((x) >= (y) ? (x) : (y))
#define MIN(x, y) ((x) <= (y) ? (x) : (y))
// =======================================
// MixerProvider functions implementation
static DeviceList deviceCache;
INT32 DAUDIO_GetDirectAudioDeviceCount() {
deviceCache.Refresh();
int count = deviceCache.GetCount();
if (count > 0) {
// add "default" device
count++;
TRACE1("DAUDIO_GetDirectAudioDeviceCount: returns %d devices\n", count);
} else {
TRACE0("DAUDIO_GetDirectAudioDeviceCount: no devices found\n");
}
return count;
}
INT32 DAUDIO_GetDirectAudioDeviceDescription(INT32 mixerIndex, DirectAudioDeviceDescription *desc) {
bool result = true;
desc->deviceID = 0;
if (mixerIndex == 0) {
// default device
strncpy(desc->name, "Default Audio Device", DAUDIO_STRING_LENGTH);
strncpy(desc->description, "Default Audio Device", DAUDIO_STRING_LENGTH);
desc->maxSimulLines = -1;
} else {
AudioDeviceID deviceID;
result = deviceCache.GetDeviceInfo(mixerIndex-1, &deviceID, DAUDIO_STRING_LENGTH,
desc->name, desc->vendor, desc->description, desc->version);
if (result) {
desc->deviceID = (INT32)deviceID;
desc->maxSimulLines = -1;
}
}
return result ? TRUE : FALSE;
}
void DAUDIO_GetFormats(INT32 mixerIndex, INT32 deviceID, int isSource, void* creator) {
TRACE3(">>DAUDIO_GetFormats mixerIndex=%d deviceID=0x%x isSource=%d\n", (int)mixerIndex, (int)deviceID, isSource);
AudioDeviceID audioDeviceID = deviceID == 0 ? GetDefaultDevice(isSource) : (AudioDeviceID)deviceID;
if (audioDeviceID == 0) {
return;
}
int totalChannels = GetChannelCount(audioDeviceID, isSource);
if (totalChannels == 0) {
TRACE0("<<DAUDIO_GetFormats, no streams!\n");
return;
}
if (isSource && totalChannels < 2) {
// report 2 channels even if only mono is supported
totalChannels = 2;
}
int channels[] = {1, 2, totalChannels};
int channelsCount = MIN(totalChannels, 3);
float hardwareSampleRate = GetSampleRate(audioDeviceID, isSource);
TRACE2(" DAUDIO_GetFormats: got %d channels, sampleRate == %f\n", totalChannels, hardwareSampleRate);
// any sample rates are supported
float sampleRate = -1;
static int sampleBits[] = {8, 16, 24};
static int sampleBitsCount = sizeof(sampleBits)/sizeof(sampleBits[0]);
// the last audio format is the default one (used by DataLine.open() if format is not specified)
// consider as default 16bit PCM stereo (mono is stereo is not supported) with the current sample rate
int defBits = 16;
int defChannels = MIN(2, channelsCount);
float defSampleRate = hardwareSampleRate;
// don't add default format is sample rate is not specified
bool addDefault = defSampleRate > 0;
// TODO: CoreAudio can handle signed/unsigned, little-endian/big-endian
// TODO: register the formats (to prevent DirectAudio software conversion) - need to fix DirectAudioDevice.createDataLineInfo
// to avoid software conversions if both signed/unsigned or big-/little-endian are supported
for (int channelIndex = 0; channelIndex < channelsCount; channelIndex++) {
for (int bitIndex = 0; bitIndex < sampleBitsCount; bitIndex++) {
int bits = sampleBits[bitIndex];
if (addDefault && bits == defBits && channels[channelIndex] != defChannels && sampleRate == defSampleRate) {
// the format is the default one, don't add it now
continue;
}
DAUDIO_AddAudioFormat(creator,
bits, // sample size in bits
-1, // frame size (auto)
channels[channelIndex], // channels
sampleRate, // sample rate
DAUDIO_PCM, // only accept PCM
bits == 8 ? FALSE : TRUE, // signed
bits == 8 ? FALSE // little-endian for 8bit
: UTIL_IsBigEndianPlatform());
}
}
// add default format
if (addDefault) {
DAUDIO_AddAudioFormat(creator,
defBits, // 16 bits
-1, // automatically calculate frame size
defChannels, // channels
defSampleRate, // sample rate
DAUDIO_PCM, // PCM
TRUE, // signed
UTIL_IsBigEndianPlatform()); // native endianess
}
TRACE0("<<DAUDIO_GetFormats\n");
}
// =======================================
// Source/Target DataLine functions implementation
// ====
/* 1writer-1reader ring buffer class with flush() support */
class RingBuffer {
public:
RingBuffer() : pBuffer(NULL), nBufferSize(0) {
pthread_mutex_init(&lockMutex, NULL);
}
~RingBuffer() {
Deallocate();
pthread_mutex_destroy(&lockMutex);
}
// extraBytes: number of additionally allocated bytes to prevent data
// overlapping when almost whole buffer is filled
// (required only if Write() can override the buffer)
bool Allocate(int requestedBufferSize, int extraBytes) {
int fullBufferSize = requestedBufferSize + extraBytes;
int powerOfTwo = 1;
while (powerOfTwo < fullBufferSize) {
powerOfTwo <<= 1;
}
pBuffer = (Byte*)malloc(powerOfTwo);
if (pBuffer == NULL) {
ERROR0("RingBuffer::Allocate: OUT OF MEMORY\n");
return false;
}
nBufferSize = requestedBufferSize;
nAllocatedBytes = powerOfTwo;
nPosMask = powerOfTwo - 1;
nWritePos = 0;
nReadPos = 0;
nFlushPos = -1;
TRACE2("RingBuffer::Allocate: OK, bufferSize=%d, allocated:%d\n", nBufferSize, nAllocatedBytes);
return true;
}
void Deallocate() {
if (pBuffer) {
free(pBuffer);
pBuffer = NULL;
nBufferSize = 0;
}
}
inline int GetBufferSize() {
return nBufferSize;
}
inline int GetAllocatedSize() {
return nAllocatedBytes;
}
// gets number of bytes available for reading
int GetValidByteCount() {
lock();
INT64 result = nWritePos - (nFlushPos >= 0 ? nFlushPos : nReadPos);
unlock();
return result > (INT64)nBufferSize ? nBufferSize : (int)result;
}
int Write(void *srcBuffer, int len, bool preventOverflow) {
lock();
TRACE2("RingBuffer::Write (%d bytes, preventOverflow=%d)\n", len, preventOverflow ? 1 : 0);
TRACE2(" writePos = %lld (%d)", (long long)nWritePos, Pos2Offset(nWritePos));
TRACE2(" readPos=%lld (%d)", (long long)nReadPos, Pos2Offset(nReadPos));
TRACE2(" flushPos=%lld (%d)\n", (long long)nFlushPos, Pos2Offset(nFlushPos));
INT64 writePos = nWritePos;
if (preventOverflow) {
INT64 avail_read = writePos - (nFlushPos >= 0 ? nFlushPos : nReadPos);
if (avail_read >= (INT64)nBufferSize) {
// no space
TRACE0(" preventOverlow: OVERFLOW => len = 0;\n");
len = 0;
} else {
int avail_write = nBufferSize - (int)avail_read;
if (len > avail_write) {
TRACE2(" preventOverlow: desrease len: %d => %d\n", len, avail_write);
len = avail_write;
}
}
}
unlock();
if (len > 0) {
write((Byte *)srcBuffer, Pos2Offset(writePos), len);
lock();
TRACE4("--RingBuffer::Write writePos: %lld (%d) => %lld, (%d)\n",
(long long)nWritePos, Pos2Offset(nWritePos), (long long)nWritePos + len, Pos2Offset(nWritePos + len));
nWritePos += len;
unlock();
}
return len;
}
int Read(void *dstBuffer, int len) {
lock();
TRACE1("RingBuffer::Read (%d bytes)\n", len);
TRACE2(" writePos = %lld (%d)", (long long)nWritePos, Pos2Offset(nWritePos));
TRACE2(" readPos=%lld (%d)", (long long)nReadPos, Pos2Offset(nReadPos));
TRACE2(" flushPos=%lld (%d)\n", (long long)nFlushPos, Pos2Offset(nFlushPos));
applyFlush();
INT64 avail_read = nWritePos - nReadPos;
// check for overflow
if (avail_read > (INT64)nBufferSize) {
nReadPos = nWritePos - nBufferSize;
avail_read = nBufferSize;
TRACE0(" OVERFLOW\n");
}
INT64 readPos = nReadPos;
unlock();
if (len > (int)avail_read) {
TRACE2(" RingBuffer::Read - don't have enough data, len: %d => %d\n", len, (int)avail_read);
len = (int)avail_read;
}
if (len > 0) {
read((Byte *)dstBuffer, Pos2Offset(readPos), len);
lock();
if (applyFlush()) {
// just got flush(), results became obsolete
TRACE0("--RingBuffer::Read, got Flush, return 0\n");
len = 0;
} else {
TRACE4("--RingBuffer::Read readPos: %lld (%d) => %lld (%d)\n",
(long long)nReadPos, Pos2Offset(nReadPos), (long long)nReadPos + len, Pos2Offset(nReadPos + len));
nReadPos += len;
}
unlock();
} else {
// underrun!
}
return len;
}
// returns number of the flushed bytes
int Flush() {
lock();
INT64 flushedBytes = nWritePos - (nFlushPos >= 0 ? nFlushPos : nReadPos);
nFlushPos = nWritePos;
unlock();
return flushedBytes > (INT64)nBufferSize ? nBufferSize : (int)flushedBytes;
}
private:
Byte *pBuffer;
int nBufferSize;
int nAllocatedBytes;
INT64 nPosMask;
pthread_mutex_t lockMutex;
volatile INT64 nWritePos;
volatile INT64 nReadPos;
// Flush() sets nFlushPos value to nWritePos;
// next Read() sets nReadPos to nFlushPos and resests nFlushPos to -1
volatile INT64 nFlushPos;
inline void lock() {
pthread_mutex_lock(&lockMutex);
}
inline void unlock() {
pthread_mutex_unlock(&lockMutex);
}
inline bool applyFlush() {
if (nFlushPos >= 0) {
nReadPos = nFlushPos;
nFlushPos = -1;
return true;
}
return false;
}
inline int Pos2Offset(INT64 pos) {
return (int)(pos & nPosMask);
}
void write(Byte *srcBuffer, int dstOffset, int len) {
int dstEndOffset = dstOffset + len;
int lenAfterWrap = dstEndOffset - nAllocatedBytes;
if (lenAfterWrap > 0) {
// dest.buffer does wrap
len = nAllocatedBytes - dstOffset;
memcpy(pBuffer+dstOffset, srcBuffer, len);
memcpy(pBuffer, srcBuffer+len, lenAfterWrap);
} else {
// dest.buffer does not wrap
memcpy(pBuffer+dstOffset, srcBuffer, len);
}
}
void read(Byte *dstBuffer, int srcOffset, int len) {
int srcEndOffset = srcOffset + len;
int lenAfterWrap = srcEndOffset - nAllocatedBytes;
if (lenAfterWrap > 0) {
// need to unwrap data
len = nAllocatedBytes - srcOffset;
memcpy(dstBuffer, pBuffer+srcOffset, len);
memcpy(dstBuffer+len, pBuffer, lenAfterWrap);
} else {
// source buffer is not wrapped
memcpy(dstBuffer, pBuffer+srcOffset, len);
}
}
};
class Resampler {
private:
enum {
kResamplerEndOfInputData = 1 // error to interrupt conversion (end of input data)
};
public:
Resampler() : converter(NULL), outBuffer(NULL) { }
~Resampler() {
if (converter != NULL) {
AudioConverterDispose(converter);
}
if (outBuffer != NULL) {
free(outBuffer);
}
}
// inFormat & outFormat must be interleaved!
bool Init(const AudioStreamBasicDescription *inFormat, const AudioStreamBasicDescription *outFormat,
int inputBufferSizeInBytes)
{
TRACE0(">>Resampler::Init\n");
TRACE0(" inFormat: ");
PrintStreamDesc(inFormat);
TRACE0(" outFormat: ");
PrintStreamDesc(outFormat);
TRACE1(" inputBufferSize: %d bytes\n", inputBufferSizeInBytes);
OSStatus err;
if ((outFormat->mFormatFlags & kAudioFormatFlagIsNonInterleaved) != 0 && outFormat->mChannelsPerFrame != 1) {
ERROR0("Resampler::Init ERROR: outFormat is non-interleaved\n");
return false;
}
if ((inFormat->mFormatFlags & kAudioFormatFlagIsNonInterleaved) != 0 && inFormat->mChannelsPerFrame != 1) {
ERROR0("Resampler::Init ERROR: inFormat is non-interleaved\n");
return false;
}
memcpy(&asbdIn, inFormat, sizeof(AudioStreamBasicDescription));
memcpy(&asbdOut, outFormat, sizeof(AudioStreamBasicDescription));
err = AudioConverterNew(inFormat, outFormat, &converter);
if (err || converter == NULL) {
OS_ERROR1(err, "Resampler::Init (AudioConverterNew), converter=%p", converter);
return false;
}
// allocate buffer for output data
int maximumInFrames = inputBufferSizeInBytes / inFormat->mBytesPerFrame;
// take into account trailingFrames
AudioConverterPrimeInfo primeInfo = {0, 0};
UInt32 sizePrime = sizeof(primeInfo);
err = AudioConverterGetProperty(converter, kAudioConverterPrimeInfo, &sizePrime, &primeInfo);
if (err) {
OS_ERROR0(err, "Resampler::Init (get kAudioConverterPrimeInfo)");
// ignore the error
} else {
// the default primeMethod is kConverterPrimeMethod_Normal, so we need only trailingFrames
maximumInFrames += primeInfo.trailingFrames;
}
float outBufferSizeInFrames = (outFormat->mSampleRate / inFormat->mSampleRate) * ((float)maximumInFrames);
// to avoid complex calculation just set outBufferSize as double of the calculated value
outBufferSize = (int)outBufferSizeInFrames * outFormat->mBytesPerFrame * 2;
// safety check - consider 256 frame as the minimum input buffer
int minOutSize = 256 * outFormat->mBytesPerFrame;
if (outBufferSize < minOutSize) {
outBufferSize = minOutSize;
}
outBuffer = malloc(outBufferSize);
if (outBuffer == NULL) {
ERROR1("Resampler::Init ERROR: malloc failed (%d bytes)\n", outBufferSize);
AudioConverterDispose(converter);
converter = NULL;
return false;
}
TRACE1(" allocated: %d bytes for output buffer\n", outBufferSize);
TRACE0("<<Resampler::Init: OK\n");
return true;
}
// returns size of the internal output buffer
int GetOutBufferSize() {
return outBufferSize;
}
// process next part of data (writes resampled data to the ringBuffer without overflow check)
int Process(void *srcBuffer, int len, RingBuffer *ringBuffer) {
int bytesWritten = 0;
TRACE2(">>Resampler::Process: %d bytes, converter = %p\n", len, converter);
if (converter == NULL) { // sanity check
bytesWritten = ringBuffer->Write(srcBuffer, len, false);
} else {
InputProcData data;
data.pThis = this;
data.data = (Byte *)srcBuffer;
data.dataSize = len;
OSStatus err;
do {
AudioBufferList abl; // by default it contains 1 AudioBuffer
abl.mNumberBuffers = 1;
abl.mBuffers[0].mNumberChannels = asbdOut.mChannelsPerFrame;
abl.mBuffers[0].mDataByteSize = outBufferSize;
abl.mBuffers[0].mData = outBuffer;
UInt32 packets = (UInt32)outBufferSize / asbdOut.mBytesPerPacket;
TRACE2(">>AudioConverterFillComplexBuffer: request %d packets, provide %d bytes buffer\n",
(int)packets, (int)abl.mBuffers[0].mDataByteSize);
err = AudioConverterFillComplexBuffer(converter, ConverterInputProc, &data, &packets, &abl, NULL);
TRACE2("<<AudioConverterFillComplexBuffer: got %d packets (%d bytes)\n",
(int)packets, (int)abl.mBuffers[0].mDataByteSize);
if (packets > 0) {
int bytesToWrite = (int)(packets * asbdOut.mBytesPerPacket);
bytesWritten += ringBuffer->Write(abl.mBuffers[0].mData, bytesToWrite, false);
}
// if outputBuffer is small to store all available frames,
// we get noErr here. In the case just continue the conversion
} while (err == noErr);
if (err != kResamplerEndOfInputData) {
// unexpected error
OS_ERROR0(err, "Resampler::Process (AudioConverterFillComplexBuffer)");
}
}
TRACE2("<<Resampler::Process: written %d bytes (converted from %d bytes)\n", bytesWritten, len);
return bytesWritten;
}
// resets internal bufferes
void Discontinue() {
TRACE0(">>Resampler::Discontinue\n");
if (converter != NULL) {
AudioConverterReset(converter);
}
TRACE0("<<Resampler::Discontinue\n");
}
private:
AudioConverterRef converter;
// buffer for output data
// note that there is no problem if the buffer is not big enough to store
// all converted data - it's only performance issue
void *outBuffer;
int outBufferSize;
AudioStreamBasicDescription asbdIn;
AudioStreamBasicDescription asbdOut;
struct InputProcData {
Resampler *pThis;
Byte *data; // data == NULL means we handle Discontinue(false)
int dataSize; // == 0 if all data was already provided to the converted of we handle Discontinue(false)
};
static OSStatus ConverterInputProc(AudioConverterRef inAudioConverter, UInt32 *ioNumberDataPackets,
AudioBufferList *ioData, AudioStreamPacketDescription **outDataPacketDescription, void *inUserData)
{
InputProcData *data = (InputProcData *)inUserData;
TRACE3(" >>ConverterInputProc: requested %d packets, data contains %d bytes (%d packets)\n",
(int)*ioNumberDataPackets, (int)data->dataSize, (int)(data->dataSize / data->pThis->asbdIn.mBytesPerPacket));
if (data->dataSize == 0) {
// already called & provided all input data
// interrupt conversion by returning error
*ioNumberDataPackets = 0;
TRACE0(" <<ConverterInputProc: returns kResamplerEndOfInputData\n");
return kResamplerEndOfInputData;
}
ioData->mNumberBuffers = 1;
ioData->mBuffers[0].mNumberChannels = data->pThis->asbdIn.mChannelsPerFrame;
ioData->mBuffers[0].mDataByteSize = data->dataSize;
ioData->mBuffers[0].mData = data->data;
*ioNumberDataPackets = data->dataSize / data->pThis->asbdIn.mBytesPerPacket;
// all data has been provided to the converter
data->dataSize = 0;
TRACE1(" <<ConverterInputProc: returns %d packets\n", (int)(*ioNumberDataPackets));
return noErr;
}
};
struct OSX_DirectAudioDevice {
AudioUnit audioUnit;
RingBuffer ringBuffer;
AudioStreamBasicDescription asbd;
// only for target lines
UInt32 inputBufferSizeInBytes;
Resampler *resampler;
// to detect discontinuity (to reset resampler)
SInt64 lastWrittenSampleTime;
OSX_DirectAudioDevice() : audioUnit(NULL), asbd(), resampler(NULL), lastWrittenSampleTime(0) {
}
~OSX_DirectAudioDevice() {
if (audioUnit) {
AudioComponentInstanceDispose(audioUnit);
}
if (resampler) {
delete resampler;
}
}
};
static AudioUnit CreateOutputUnit(AudioDeviceID deviceID, int isSource)
{
OSStatus err;
AudioUnit unit;
AudioComponentDescription desc;
desc.componentType = kAudioUnitType_Output;
desc.componentSubType = (deviceID == 0 && isSource) ? kAudioUnitSubType_DefaultOutput : kAudioUnitSubType_HALOutput;
desc.componentManufacturer = kAudioUnitManufacturer_Apple;
desc.componentFlags = 0;
desc.componentFlagsMask = 0;
AudioComponent comp = AudioComponentFindNext(NULL, &desc);
err = AudioComponentInstanceNew(comp, &unit);
if (err) {
OS_ERROR0(err, "CreateOutputUnit:OpenAComponent");
return NULL;
}
if (!isSource) {
int enableIO = 0;
err = AudioUnitSetProperty(unit, kAudioOutputUnitProperty_EnableIO, kAudioUnitScope_Output,
0, &enableIO, sizeof(enableIO));
if (err) {
OS_ERROR0(err, "SetProperty (output EnableIO)");
}
enableIO = 1;
err = AudioUnitSetProperty(unit, kAudioOutputUnitProperty_EnableIO, kAudioUnitScope_Input,
1, &enableIO, sizeof(enableIO));
if (err) {
OS_ERROR0(err, "SetProperty (input EnableIO)");
}
if (!deviceID) {
// get real AudioDeviceID for default input device (macosx current input device)
deviceID = GetDefaultDevice(isSource);
if (!deviceID) {
AudioComponentInstanceDispose(unit);
return NULL;
}
}
}
if (deviceID) {
err = AudioUnitSetProperty(unit, kAudioOutputUnitProperty_CurrentDevice, kAudioUnitScope_Global,
0, &deviceID, sizeof(deviceID));
if (err) {
OS_ERROR0(err, "SetProperty (CurrentDevice)");
AudioComponentInstanceDispose(unit);
return NULL;
}
}
return unit;
}
static OSStatus OutputCallback(void *inRefCon,
AudioUnitRenderActionFlags *ioActionFlags,
const AudioTimeStamp *inTimeStamp,
UInt32 inBusNumber,
UInt32 inNumberFrames,
AudioBufferList *ioData)
{
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)inRefCon;
int nchannels = ioData->mNumberBuffers; // should be always == 1 (interleaved channels)
AudioBuffer *audioBuffer = ioData->mBuffers;
TRACE3(">>OutputCallback: busNum=%d, requested %d frames (%d bytes)\n",
(int)inBusNumber, (int)inNumberFrames, (int)(inNumberFrames * device->asbd.mBytesPerFrame));
TRACE3(" abl: %d buffers, buffer[0].channels=%d, buffer.size=%d\n",
nchannels, (int)audioBuffer->mNumberChannels, (int)audioBuffer->mDataByteSize);
int bytesToRead = inNumberFrames * device->asbd.mBytesPerFrame;
if (bytesToRead > (int)audioBuffer->mDataByteSize) {
TRACE0("--OutputCallback: !!! audioBuffer IS TOO SMALL!!!\n");
bytesToRead = audioBuffer->mDataByteSize / device->asbd.mBytesPerFrame * device->asbd.mBytesPerFrame;
}
int bytesRead = device->ringBuffer.Read(audioBuffer->mData, bytesToRead);
if (bytesRead < bytesToRead) {
// no enough data (underrun)
TRACE2("--OutputCallback: !!! UNDERRUN (read %d bytes of %d)!!!\n", bytesRead, bytesToRead);
// silence the rest
memset((Byte*)audioBuffer->mData + bytesRead, 0, bytesToRead-bytesRead);
bytesRead = bytesToRead;
}
audioBuffer->mDataByteSize = (UInt32)bytesRead;
// SAFETY: set mDataByteSize for all other AudioBuffer in the AudioBufferList to zero
while (--nchannels > 0) {
audioBuffer++;
audioBuffer->mDataByteSize = 0;
}
TRACE1("<<OutputCallback (returns %d)\n", bytesRead);
return noErr;
}
static OSStatus InputCallback(void *inRefCon,
AudioUnitRenderActionFlags *ioActionFlags,
const AudioTimeStamp *inTimeStamp,
UInt32 inBusNumber,
UInt32 inNumberFrames,
AudioBufferList *ioData)
{
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)inRefCon;
TRACE4(">>InputCallback: busNum=%d, timeStamp=%lld, %d frames (%d bytes)\n",
(int)inBusNumber, (long long)inTimeStamp->mSampleTime, (int)inNumberFrames, (int)(inNumberFrames * device->asbd.mBytesPerFrame));
AudioBufferList abl; // by default it contains 1 AudioBuffer
abl.mNumberBuffers = 1;
abl.mBuffers[0].mNumberChannels = device->asbd.mChannelsPerFrame;
abl.mBuffers[0].mDataByteSize = device->inputBufferSizeInBytes; // assume this is == (inNumberFrames * device->asbd.mBytesPerFrame)
abl.mBuffers[0].mData = NULL; // request for the audioUnit's buffer
OSStatus err = AudioUnitRender(device->audioUnit, ioActionFlags, inTimeStamp, inBusNumber, inNumberFrames, &abl);
if (err) {
OS_ERROR0(err, "<<InputCallback: AudioUnitRender");
} else {
if (device->resampler != NULL) {
// test for discontinuity
// AUHAL starts timestamps at zero, so test if the current timestamp less then the last written
SInt64 sampleTime = inTimeStamp->mSampleTime;
if (sampleTime < device->lastWrittenSampleTime) {
// discontinuity, reset the resampler
TRACE2(" InputCallback (RESAMPLED), DISCONTINUITY (%f -> %f)\n",
(float)device->lastWrittenSampleTime, (float)sampleTime);
device->resampler->Discontinue();
} else {
TRACE2(" InputCallback (RESAMPLED), continuous: lastWrittenSampleTime = %f, sampleTime=%f\n",
(float)device->lastWrittenSampleTime, (float)sampleTime);
}
device->lastWrittenSampleTime = sampleTime + inNumberFrames;
int bytesWritten = device->resampler->Process(abl.mBuffers[0].mData, (int)abl.mBuffers[0].mDataByteSize, &device->ringBuffer);
TRACE2("<<InputCallback (RESAMPLED, saved %d bytes of %d)\n", bytesWritten, (int)abl.mBuffers[0].mDataByteSize);
} else {
int bytesWritten = device->ringBuffer.Write(abl.mBuffers[0].mData, (int)abl.mBuffers[0].mDataByteSize, false);
TRACE2("<<InputCallback (saved %d bytes of %d)\n", bytesWritten, (int)abl.mBuffers[0].mDataByteSize);
}
}
return noErr;
}
static void FillASBDForNonInterleavedPCM(AudioStreamBasicDescription& asbd,
float sampleRate, int channels, int sampleSizeInBits, bool isFloat, int isSigned, bool isBigEndian)
{
// FillOutASBDForLPCM cannot produce unsigned integer format
asbd.mSampleRate = sampleRate;
asbd.mFormatID = kAudioFormatLinearPCM;
asbd.mFormatFlags = (isFloat ? kAudioFormatFlagIsFloat : (isSigned ? kAudioFormatFlagIsSignedInteger : 0))
| (isBigEndian ? (kAudioFormatFlagIsBigEndian) : 0)
| kAudioFormatFlagIsPacked;
asbd.mBytesPerPacket = channels * ((sampleSizeInBits + 7) / 8);
asbd.mFramesPerPacket = 1;
asbd.mBytesPerFrame = asbd.mBytesPerPacket;
asbd.mChannelsPerFrame = channels;
asbd.mBitsPerChannel = sampleSizeInBits;
}
void* DAUDIO_Open(INT32 mixerIndex, INT32 deviceID, int isSource,
int encoding, float sampleRate, int sampleSizeInBits,
int frameSize, int channels,
int isSigned, int isBigEndian, int bufferSizeInBytes)
{
TRACE3(">>DAUDIO_Open: mixerIndex=%d deviceID=0x%x isSource=%d\n", (int)mixerIndex, (unsigned int)deviceID, isSource);
TRACE3(" sampleRate=%d sampleSizeInBits=%d channels=%d\n", (int)sampleRate, sampleSizeInBits, channels);
#ifdef USE_TRACE
{
AudioDeviceID audioDeviceID = deviceID;
if (audioDeviceID == 0) {
// default device
audioDeviceID = GetDefaultDevice(isSource);
}
char name[256];
OSStatus err = GetAudioObjectProperty(audioDeviceID, kAudioUnitScope_Global, kAudioDevicePropertyDeviceName, 256, &name, 0);
if (err != noErr) {
OS_ERROR1(err, " audioDeviceID=0x%x, name is N/A:", (int)audioDeviceID);
} else {
TRACE2(" audioDeviceID=0x%x, name=%s\n", (int)audioDeviceID, name);
}
}
#endif
if (encoding != DAUDIO_PCM) {
ERROR1("<<DAUDIO_Open: ERROR: unsupported encoding (%d)\n", encoding);
return NULL;
}
if (channels <= 0) {
ERROR1("<<DAUDIO_Open: ERROR: Invalid number of channels=%d!\n", channels);
return NULL;
}
OSX_DirectAudioDevice *device = new OSX_DirectAudioDevice();
AudioUnitScope scope = isSource ? kAudioUnitScope_Input : kAudioUnitScope_Output;
int element = isSource ? 0 : 1;
OSStatus err = noErr;
int extraBufferBytes = 0;
device->audioUnit = CreateOutputUnit(deviceID, isSource);
if (!device->audioUnit) {
delete device;
return NULL;
}
if (!isSource) {
AudioDeviceID actualDeviceID = deviceID != 0 ? deviceID : GetDefaultDevice(isSource);
float hardwareSampleRate = GetSampleRate(actualDeviceID, isSource);
TRACE2("--DAUDIO_Open: sampleRate = %f, hardwareSampleRate=%f\n", sampleRate, hardwareSampleRate);
if (fabs(sampleRate - hardwareSampleRate) > 1) {
device->resampler = new Resampler();
// request HAL for Float32 with native endianess
FillASBDForNonInterleavedPCM(device->asbd, hardwareSampleRate, channels, 32, true, false, kAudioFormatFlagsNativeEndian != 0);
} else {
sampleRate = hardwareSampleRate; // in case sample rates are not exactly equal
}
}
if (device->resampler == NULL) {
// no resampling, request HAL for the requested format
FillASBDForNonInterleavedPCM(device->asbd, sampleRate, channels, sampleSizeInBits, false, isSigned, isBigEndian);
}
err = AudioUnitSetProperty(device->audioUnit, kAudioUnitProperty_StreamFormat, scope, element, &device->asbd, sizeof(device->asbd));
if (err) {
OS_ERROR0(err, "<<DAUDIO_Open set StreamFormat");
delete device;
return NULL;
}
AURenderCallbackStruct output;
output.inputProc = isSource ? OutputCallback : InputCallback;
output.inputProcRefCon = device;
err = AudioUnitSetProperty(device->audioUnit,
isSource
? (AudioUnitPropertyID)kAudioUnitProperty_SetRenderCallback
: (AudioUnitPropertyID)kAudioOutputUnitProperty_SetInputCallback,
kAudioUnitScope_Global, 0, &output, sizeof(output));
if (err) {
OS_ERROR0(err, "<<DAUDIO_Open set RenderCallback");
delete device;
return NULL;
}
err = AudioUnitInitialize(device->audioUnit);
if (err) {
OS_ERROR0(err, "<<DAUDIO_Open UnitInitialize");
delete device;
return NULL;
}
if (!isSource) {
// for target lines we need extra bytes in the ringBuffer
// to prevent collisions when InputCallback overrides data on overflow
UInt32 size;
OSStatus err;
size = sizeof(device->inputBufferSizeInBytes);
err = AudioUnitGetProperty(device->audioUnit, kAudioDevicePropertyBufferFrameSize, kAudioUnitScope_Global,
0, &device->inputBufferSizeInBytes, &size);
if (err) {
OS_ERROR0(err, "<<DAUDIO_Open (TargetDataLine)GetBufferSize\n");
delete device;
return NULL;
}
device->inputBufferSizeInBytes *= device->asbd.mBytesPerFrame; // convert frames to bytes
extraBufferBytes = (int)device->inputBufferSizeInBytes;
}
if (device->resampler != NULL) {
// resampler output format is a user requested format (== ringBuffer format)
AudioStreamBasicDescription asbdOut; // ringBuffer format
FillASBDForNonInterleavedPCM(asbdOut, sampleRate, channels, sampleSizeInBits, false, isSigned, isBigEndian);
// set resampler input buffer size to the HAL buffer size
if (!device->resampler->Init(&device->asbd, &asbdOut, (int)device->inputBufferSizeInBytes)) {
ERROR0("<<DAUDIO_Open: resampler.Init() FAILED.\n");
delete device;
return NULL;
}
// extra bytes in the ringBuffer (extraBufferBytes) should be equal resampler output buffer size
extraBufferBytes = device->resampler->GetOutBufferSize();
}
if (!device->ringBuffer.Allocate(bufferSizeInBytes, extraBufferBytes)) {
ERROR0("<<DAUDIO_Open: Ring buffer allocation error\n");
delete device;
return NULL;
}
TRACE0("<<DAUDIO_Open: OK\n");
return device;
}
int DAUDIO_Start(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
TRACE0("DAUDIO_Start\n");
OSStatus err = AudioOutputUnitStart(device->audioUnit);
if (err != noErr) {
OS_ERROR0(err, "DAUDIO_Start");
}
return err == noErr ? TRUE : FALSE;
}
int DAUDIO_Stop(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
TRACE0("DAUDIO_Stop\n");
OSStatus err = AudioOutputUnitStop(device->audioUnit);
return err == noErr ? TRUE : FALSE;
}
void DAUDIO_Close(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
TRACE0("DAUDIO_Close\n");
delete device;
}
int DAUDIO_Write(void* id, char* data, int byteSize) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
TRACE1(">>DAUDIO_Write: %d bytes to write\n", byteSize);
int result = device->ringBuffer.Write(data, byteSize, true);
TRACE1("<<DAUDIO_Write: %d bytes written\n", result);
return result;
}
int DAUDIO_Read(void* id, char* data, int byteSize) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
TRACE1(">>DAUDIO_Read: %d bytes to read\n", byteSize);
int result = device->ringBuffer.Read(data, byteSize);
TRACE1("<<DAUDIO_Read: %d bytes has been read\n", result);
return result;
}
int DAUDIO_GetBufferSize(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
int bufferSizeInBytes = device->ringBuffer.GetBufferSize();
TRACE1("DAUDIO_GetBufferSize returns %d\n", bufferSizeInBytes);
return bufferSizeInBytes;
}
int DAUDIO_StillDraining(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
int draining = device->ringBuffer.GetValidByteCount() > 0 ? TRUE : FALSE;
TRACE1("DAUDIO_StillDraining returns %d\n", draining);
return draining;
}
int DAUDIO_Flush(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
TRACE0("DAUDIO_Flush\n");
device->ringBuffer.Flush();
return TRUE;
}
int DAUDIO_GetAvailable(void* id, int isSource) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
int bytesInBuffer = device->ringBuffer.GetValidByteCount();
if (isSource) {
return device->ringBuffer.GetBufferSize() - bytesInBuffer;
} else {
return bytesInBuffer;
}
}
INT64 DAUDIO_GetBytePosition(void* id, int isSource, INT64 javaBytePos) {
OSX_DirectAudioDevice *device = (OSX_DirectAudioDevice*)id;
INT64 position;
if (isSource) {
position = javaBytePos - device->ringBuffer.GetValidByteCount();
} else {
position = javaBytePos + device->ringBuffer.GetValidByteCount();
}
TRACE2("DAUDIO_GetBytePosition returns %lld (javaBytePos = %lld)\n", (long long)position, (long long)javaBytePos);
return position;
}
void DAUDIO_SetBytePosition(void* id, int isSource, INT64 javaBytePos) {
// no need javaBytePos (it's available in DAUDIO_GetBytePosition)
}
int DAUDIO_RequiresServicing(void* id, int isSource) {
return FALSE;
}
void DAUDIO_Service(void* id, int isSource) {
// unreachable
}
#endif // USE_DAUDIO == TRUE