Re: custom AudioUnits and processing iTunes library on iPhone
Re: custom AudioUnits and processing iTunes library on iPhone
- Subject: Re: custom AudioUnits and processing iTunes library on iPhone
- From: Kevin Dixon <email@hidden>
- Date: Thu, 19 Apr 2012 15:25:46 -0700
Maybe a silly question, but using these or similar APIs and the same
design approach, could one also play a video, while processing the
audio stream?
-Kevin
On Tue, Apr 3, 2012 at 1:21 PM, Bill Phillips <email@hidden> wrote:
> On Tue, Apr 3, 2012 at 3:02 PM, Chris Adamson <email@hidden> wrote:
>> You can get sample-level access to the user's music by using the MPMediaItemPropertyAssetURL, using it to open the song as an AVURLAsset with the AV Foundation framework, then using the AVAssetReader API to read samples. Once you have samples, you could run them through an AUGraph and perform whatever processing you like. There are certain threading and buffering considerations that make this harder than you might be ready for, to say nothing of having to juggle three media APIs, but that should be the basic idea.
>
> I found that my easy, functional AVURLAsset solution had issues with
> decoding assets while in the background. It seems to use mediaserverd
> to do any decoding work, and mediaserverd has its own ideas about
> managing the hardware decoder. My solution was to use AVURLAsset to
> get the encoded data, then decode it using an AudioQueue.
>
> --
> Bill
_______________________________________________
Do not post admin requests to the list. They will be ignored.
Coreaudio-api mailing list (email@hidden)
Help/Unsubscribe/Update your Subscription:
This email sent to email@hidden