Re: How to capture audio generated by a render callback, iOS
Re: How to capture audio generated by a render callback, iOS
- Subject: Re: How to capture audio generated by a render callback, iOS
- From: Joel Reymont <email@hidden>
- Date: Wed, 05 Dec 2012 16:55:16 +0000
Why not change the file ASBD to use non-interleaved data?
Just add the flag and it should work.
--
http://www.linkedin.com/in/joelreymont
On Wednesday, December 5, 2012 at 4:41 PM, Tim Kemp wrote:
Ah. Thanks. I've posted a more recent follow up a couple of hours ago after I started to come to your conclusion.
Currently I'm generating non-interleaved audio (ie, two buffers), which is working fine with the RemoteIO unit. What's the best way for me to convert that to interleaved data for Extended Audio File Services? I don't want to be mallocing a new AudioBuffer in my render callback.
Could I instead have my RemoteIO use interleaved data? If I go that route, I'm not sure how I would set up the ASBD any differently.
Thanks
On 5 Dec 2012, at 11:32, Joel Reymont wrote:
Stack trace says your buffer list is bogus.
You need to provide a buffer list with at least one buffer with interleaved samples.
Your mData would have a layout like this [LRLRLR...] where L and R are SInt16 each and LR constituted one frame.
The number of frames in your mData would match numFrames passed to ExtAudioFileWriteAsync.
On Wednesday, December 5, 2012 at 1:53 AM, Tim Kemp wrote:
_______________________________________________
Do not post admin requests to the list. They will be ignored.
Coreaudio-api mailing list (email@hidden)
Help/Unsubscribe/Update your Subscription:
This email sent to email@hidden