在iOS上使用OpenAL进行声音捕获

时间:2012-05-25 15:13:07

标签: ios openal

我正在尝试使用OpenAL在iOS上进行声音捕获(我正在编写一个跨平台库,这就是为什么我要避免使用iOS特定的方式来录制声音)。 开箱即用的OpenAL捕获功能不起作用,但存在一个已知的解决方法:open an output context before starting capture。这个解决方案适用于iOS 5.0。

但是在iOS 5.1.1上,解决方法仅对我尝试录制的第一个示例有帮助。 (我在开始捕获之前将AudioSession切换到PlayAndRecord并打开默认输出设备。录制完样本后,我关闭设备并将会话切换回原来的状态。) 对于第二个示例,重新打开输出上下文没有帮助,也没有捕获声音。

有没有一种已知的方法可以解决这个问题?

// Here's what I do before starting the recording
oldAudioSessionCategory = [audioSession category];
[audioSession setCategory:AVAudioSessionCategoryPlayAndRecord error:nil];
[audioSession setActive:YES error:nil];
// We need to have an active context. If there is none, create one.
if (!alcGetCurrentContext()) {
    outputDevice = alcOpenDevice(NULL);
    outputContext = alcCreateContext(outputDevice, NULL);
    alcMakeContextCurrent(outputContext);
}

// Capture itself
inputDevice = alcCaptureOpenDevice(NULL, frequency, FORMAT, bufferSize);
....
alcCaptureCloseDevice(inputDevice);

// Restoring the audio state to whatever it had been before capture
if (outputContext) {
    alcDestroyContext(outputContext);
    alcCloseDevice(outputDevice);
}
[[AVAudioSession sharedInstance] setCategory:oldAudioSessionCategory 
                                 error:nil];

2 个答案:

答案 0 :(得分:3)

以下是我用来模拟捕获扩展的代码。 一些评论:

  1. 在整个项目中,OpenKD用于例如线程化原语。您可能需要更换这些电话。
  2. 我必须在开始捕获时对抗延迟。结果,我不断地继续读取声音输入并在不需要时扔掉它。 (例如,提出了这样的解决方案,here。)这反过来需要捕获有效的通知,以释放对麦克风的控制。您可能想也可能不想使用这样的污泥。
  3. 我必须定义一个单独的函数alcGetIntegerv(device, ALC_CAPTURE_SAMPLES, 1, &res)
  4. ,而不是alcGetAvailableSamples

    简而言之,此代码不太可能在您的项目中按原样使用,但希望您可以 根据您的需要调整它。

    #include <stdbool.h>
    #include <stddef.h>
    #include <stdint.h>
    #include <KD/kd.h>
    #include <AL/al.h>
    #include <AL/alc.h>
    
    #include <AudioToolbox/AudioToolbox.h>
    #import <Foundation/Foundation.h>
    #import <UIKit/UIKit.h>
    
    #include "KD/kdext.h"
    
    struct InputDeviceData {
        int id;
        KDThreadMutex *mutex;
        AudioUnit audioUnit;
        int nChannels;
        int frequency;
        ALCenum format;
        int sampleSize;
        uint8_t *buf;
        size_t bufSize;    // in bytes
        size_t bufFilledBytes;  // in bytes
        bool started;
    };
    
    static struct InputDeviceData *cachedInData = NULL;
    
    static OSStatus renderCallback (void                        *inRefCon,
                                    AudioUnitRenderActionFlags  *ioActionFlags,
                                    const AudioTimeStamp        *inTimeStamp,
                                    UInt32                      inBusNumber,
                                    UInt32                      inNumberFrames,
                                    AudioBufferList             *ioData);
    static AudioUnit getAudioUnit();
    static void setupNotifications();
    static void destroyCachedInData();
    static struct InputDeviceData *setupCachedInData(AudioUnit audioUnit, ALCuint frequency, ALCenum format, ALCsizei bufferSizeInSamples);
    static struct InputDeviceData *getInputDeviceData(AudioUnit audioUnit, ALCuint frequency, ALCenum format, ALCsizei bufferSizeInSamples);
    
    /** I only have to use NSNotificationCenter instead of CFNotificationCenter
     *  because there is no published name for WillResignActive/WillBecomeActive
     *  notifications in CoreFoundation.
     */
    @interface ALCNotificationObserver : NSObject
    - (void)onResignActive;
    @end
    @implementation ALCNotificationObserver
    - (void)onResignActive {
        destroyCachedInData();
    }
    @end
    
    static void setupNotifications() {
        static ALCNotificationObserver *observer = NULL;
        if (!observer) {
            observer = [[ALCNotificationObserver alloc] init];
            [[NSNotificationCenter defaultCenter] addObserver:observer selector:@selector(onResignActive) name:UIApplicationWillResignActiveNotification object:nil];
        }
    }
    
    static OSStatus renderCallback (void                        *inRefCon,
                                    AudioUnitRenderActionFlags  *ioActionFlags,
                                    const AudioTimeStamp        *inTimeStamp,
                                    UInt32                      inBusNumber,
                                    UInt32                      inNumberFrames,
                                    AudioBufferList             *ioData) {
        struct InputDeviceData *inData = (struct InputDeviceData*)inRefCon;
    
        kdThreadMutexLock(inData->mutex);
        size_t bytesToRender = inNumberFrames * inData->sampleSize;
        if (bytesToRender + inData->bufFilledBytes <= inData->bufSize) {
            OSStatus status;
            struct AudioBufferList audioBufferList; // 1 buffer is declared inside the structure itself.
            audioBufferList.mNumberBuffers = 1;
            audioBufferList.mBuffers[0].mNumberChannels = inData->nChannels;
            audioBufferList.mBuffers[0].mDataByteSize = bytesToRender;
            audioBufferList.mBuffers[0].mData = inData->buf + inData->bufFilledBytes;
            status = AudioUnitRender(inData->audioUnit, 
                                     ioActionFlags, 
                                     inTimeStamp, 
                                     inBusNumber, 
                                     inNumberFrames, 
                                     &audioBufferList);
            if (inData->started) {
                inData->bufFilledBytes += bytesToRender;
            }
        } else {
            kdLogFormatMessage("%s: buffer overflow", __FUNCTION__);
        }
        kdThreadMutexUnlock(inData->mutex);
    
        return 0;
    }
    
    static AudioUnit getAudioUnit() {
        static AudioUnit audioUnit = NULL;
    
        if (!audioUnit) {
            AudioComponentDescription ioUnitDescription;
    
            ioUnitDescription.componentType          = kAudioUnitType_Output;
            ioUnitDescription.componentSubType       = kAudioUnitSubType_VoiceProcessingIO;
            ioUnitDescription.componentManufacturer  = kAudioUnitManufacturer_Apple;
            ioUnitDescription.componentFlags         = 0;
            ioUnitDescription.componentFlagsMask     = 0;
    
            AudioComponent foundIoUnitReference = AudioComponentFindNext(NULL,
                                                                         &ioUnitDescription);
            AudioComponentInstanceNew(foundIoUnitReference,
                                      &audioUnit);
    
            if (audioUnit == NULL) {
                kdLogMessage("Could not obtain AudioUnit");
            }
        }
    
        return audioUnit;
    }
    
    static void destroyCachedInData() {
        OSStatus status;
        if (cachedInData) {
            status = AudioOutputUnitStop(cachedInData->audioUnit);
            status = AudioUnitUninitialize(cachedInData->audioUnit);
            free(cachedInData->buf);
            kdThreadMutexFree(cachedInData->mutex);
            free(cachedInData);
            cachedInData = NULL;
        }
    }
    
    static struct InputDeviceData *setupCachedInData(AudioUnit audioUnit, ALCuint frequency, ALCenum format, ALCsizei bufferSizeInSamples) {
        static int idCount = 0;
        OSStatus status;
        int bytesPerFrame = (format == AL_FORMAT_MONO8) ? 1 :
                            (format == AL_FORMAT_MONO16) ? 2 :
                            (format == AL_FORMAT_STEREO8) ? 2 :
                            (format == AL_FORMAT_STEREO16) ? 4 : -1;
        int channelsPerFrame = (format == AL_FORMAT_MONO8) ? 1 :
                               (format == AL_FORMAT_MONO16) ? 1 :
                               (format == AL_FORMAT_STEREO8) ? 2 :
                               (format == AL_FORMAT_STEREO16) ? 2 : -1;
        int bitsPerChannel = (format == AL_FORMAT_MONO8) ? 8 :
                             (format == AL_FORMAT_MONO16) ? 16 :
                             (format == AL_FORMAT_STEREO8) ? 8 :
                             (format == AL_FORMAT_STEREO16) ? 16 : -1;
    
        cachedInData = malloc(sizeof(struct InputDeviceData));
        cachedInData->id = ++idCount;
        cachedInData->format = format;
        cachedInData->frequency = frequency;
        cachedInData->mutex = kdThreadMutexCreate(NULL);
        cachedInData->audioUnit = audioUnit;
        cachedInData->nChannels = channelsPerFrame;
        cachedInData->sampleSize = bytesPerFrame;
        cachedInData->bufSize = bufferSizeInSamples * bytesPerFrame;
        cachedInData->buf = malloc(cachedInData->bufSize);
        cachedInData->bufFilledBytes = 0;
        cachedInData->started = FALSE;
    
        UInt32 enableOutput        = 1;    // to enable output
        status = AudioUnitSetProperty(audioUnit,
                                      kAudioOutputUnitProperty_EnableIO,
                                      kAudioUnitScope_Input,
                                      1,
                                      &enableOutput, sizeof(enableOutput));
    
        struct AudioStreamBasicDescription basicDescription;
        basicDescription.mSampleRate = (Float64)frequency;
        basicDescription.mFormatID = kAudioFormatLinearPCM;
        basicDescription.mFormatFlags = kAudioFormatFlagIsSignedInteger | kAudioFormatFlagIsPacked;
        basicDescription.mBytesPerPacket = bytesPerFrame;
        basicDescription.mFramesPerPacket = 1;
        basicDescription.mBytesPerFrame = bytesPerFrame;
        basicDescription.mChannelsPerFrame = channelsPerFrame;
        basicDescription.mBitsPerChannel = bitsPerChannel;
        basicDescription.mReserved = 0;
    
        status = AudioUnitSetProperty(audioUnit, 
                                      kAudioUnitProperty_StreamFormat, // property key 
                                      kAudioUnitScope_Output,        // scope
                                      1,                             // 1 is output
                                      &basicDescription, sizeof(basicDescription));      // value
    
        AURenderCallbackStruct renderCallbackStruct;
        renderCallbackStruct.inputProc = renderCallback;
        renderCallbackStruct.inputProcRefCon = cachedInData;
        status = AudioUnitSetProperty(audioUnit, 
                                      kAudioOutputUnitProperty_SetInputCallback, // property key 
                                      kAudioUnitScope_Output,        // scope
                                      1,                             // 1 is output
                                      &renderCallbackStruct, sizeof(renderCallbackStruct));      // value
    
        status = AudioOutputUnitStart(cachedInData->audioUnit);
    
        return cachedInData;
    }
    
    static struct InputDeviceData *getInputDeviceData(AudioUnit audioUnit, ALCuint frequency, ALCenum format, ALCsizei bufferSizeInSamples) {
        if (cachedInData && 
            (cachedInData->frequency != frequency ||
             cachedInData->format != format ||
             cachedInData->bufSize / cachedInData->sampleSize != bufferSizeInSamples)) {
                kdAssert(!cachedInData->started);
                destroyCachedInData();
            }
        if (!cachedInData) {
            setupCachedInData(audioUnit, frequency, format, bufferSizeInSamples);
            setupNotifications();
        }
    
        return cachedInData;
    }
    
    
    ALC_API ALCdevice* ALC_APIENTRY alcCaptureOpenDevice(const ALCchar *devicename, ALCuint frequency, ALCenum format, ALCsizei buffersizeInSamples) {
        kdAssert(devicename == NULL);    
    
        AudioUnit audioUnit = getAudioUnit();
        struct InputDeviceData *res = getInputDeviceData(audioUnit, frequency, format, buffersizeInSamples);
        return (ALCdevice*)res->id;
    }
    
    ALC_API ALCboolean ALC_APIENTRY alcCaptureCloseDevice(ALCdevice *device) {
        alcCaptureStop(device);
        return true;
    }
    
    ALC_API void ALC_APIENTRY alcCaptureStart(ALCdevice *device) {
        if (!cachedInData || (int)device != cachedInData->id) {
            // may happen after the app loses and regains active status.
            kdLogFormatMessage("Attempt to start a stale AL capture device");
            return;
        }
        cachedInData->started = TRUE;
    }
    
    ALC_API void ALC_APIENTRY alcCaptureStop(ALCdevice *device) {
        if (!cachedInData || (int)device != cachedInData->id) {
            // may happen after the app loses and regains active status.
            kdLogFormatMessage("Attempt to stop a stale AL capture device");
            return;
        }
        cachedInData->started = FALSE;
    }
    
    ALC_API ALCint ALC_APIENTRY alcGetAvailableSamples(ALCdevice *device) {
        if (!cachedInData || (int)device != cachedInData->id) {
            // may happen after the app loses and regains active status.
            kdLogFormatMessage("Attempt to get sample count from a stale AL capture device");
            return 0;
        }
        ALCint res;
        kdThreadMutexLock(cachedInData->mutex);
        res = cachedInData->bufFilledBytes / cachedInData->sampleSize;
        kdThreadMutexUnlock(cachedInData->mutex);
        return res;
    }
    
    ALC_API void ALC_APIENTRY alcCaptureSamples(ALCdevice *device, ALCvoid *buffer, ALCsizei samples) {    
        if (!cachedInData || (int)device != cachedInData->id) {
            // may happen after the app loses and regains active status.
            kdLogFormatMessage("Attempt to get samples from a stale AL capture device");
            return;
        }
        size_t bytesToCapture = samples * cachedInData->sampleSize;
        kdAssert(cachedInData->started);
        kdAssert(bytesToCapture <= cachedInData->bufFilledBytes);
    
        kdThreadMutexLock(cachedInData->mutex);
        memcpy(buffer, cachedInData->buf, bytesToCapture);
        memmove(cachedInData->buf, cachedInData->buf + bytesToCapture, cachedInData->bufFilledBytes - bytesToCapture);
        cachedInData->bufFilledBytes -= bytesToCapture;
        kdThreadMutexUnlock(cachedInData->mutex);
    }
    

答案 1 :(得分:1)

我找到了一种让Apple的OpenAL工作的方法。 在我的原始代码段中,您需要在alcMakeContextCurrent(NULL)之前调用alcDestroyContext(outputContext)