Re: Advice with timing & Audio Units (iOS)
Re: Advice with timing & Audio Units (iOS)
- Subject: Re: Advice with timing & Audio Units (iOS)
- From: Kevin Dixon <email@hidden>
- Date: Wed, 13 Aug 2014 12:41:15 -0700
It seems you may be more interested in the property
kAudioUnitProperty_Latency. As far as I know, time stamps should be
consistent across channels, etc, as the entire Audio Unit Graph should
be representing the same point in time.
Given the Latency property, you should be able to compensate by
delaying other instruments to match the greatest reported latency.
-Kevin
On Wed, Aug 13, 2014 at 12:23 PM, IDAS <email@hidden> wrote:
> I’ve recently tackled Audio Unit programming and absorbed the documentation
> on Apples website. Whilst I feel I have a good grasp of everything
> conceptually, I wanted to check my approach to writing a sequencing app on
> iOS.
>
> So my question is around the timing required to drive the sequencer. When
> recording or playing back audio (& MIDI), I want to understand the process
> of synchronising the render callbacks in the Audio Units to my master clock
> on the device.
>
>
> Audio Units have their own AudioTimeStamp and I wanted to understand more
> about how this is sourced and how it can also be manipulated to hide
> latency.
>
>
> Programatically, am I thinking about this correctly?
>
>
> Thanks.
>
>
> _______________________________________________
> Do not post admin requests to the list. They will be ignored.
> Coreaudio-api mailing list (email@hidden)
> Help/Unsubscribe/Update your Subscription:
>
> This email sent to email@hidden
_______________________________________________
Do not post admin requests to the list. They will be ignored.
Coreaudio-api mailing list (email@hidden)
Help/Unsubscribe/Update your Subscription:
This email sent to email@hidden