Repository URL to install this package:
|
Version:
3.2.0 ▾
|
{!
@file AUComponent.h
@framework AudioUnit.framework
@copyright (c) 2002-2015 Apple, Inc. All rights reserved.
@brief C interfaces for working with Audio Units.
@discussion
An audio unit is a plugin that can be loaded into an application's process and used to process
or generate audio. An audio unit is an AudioComponent type and so the AudioComponent APIs are
used to find specific types of audio units, open and close them.
Audio units use a general notion of description to specify an unique instance. The Type is the
general category of an audio unit. The SubType is an unique identifier specified by the
Manufacturer (provider) of the audio unit. The IDs that are used for Type are specified by
Apple, the Manufacturer ID should be specified by an unique identifier
(as registered with apple). See AudioComponentDescription.
Audio unit types are of the following (see below for more information)
kAudioUnitType_Output = 'auou',
kAudioUnitType_MusicDevice = 'aumu',
kAudioUnitType_MusicEffect = 'aumf',
kAudioUnitType_FormatConverter = 'aufc',
kAudioUnitType_Effect = 'aufx',
kAudioUnitType_Mixer = 'aumx',
kAudioUnitType_Panner = 'aupn',
kAudioUnitType_OfflineEffect = 'auol',
kAudioUnitType_Generator = 'augn',
An audio unit's general operations are:
- Open an audio unit (AudioComponentInstanceNew)
- Configure it based on the context (AudioUnitSetProperty)
- Initialize the audio unit (AudioUnitInitialize)
- at this point the audio unit is in a state where it can render audio
- Render audio (AudioUnitRender)
An important part of a render operation for an audio unit is to manipulate the various controls
that the unit provides to change the render effects; for instance to change the decay time of
a reverb, the cut off frequency of a filter, etc. These are called parameters, and
AudioUnitGetParameter and AudioUnitSetParameter are used to interact with these.
If any reconfiguration of the audio unit is required, then:
- Uninitialize (AudioUnitUninitialize)
- Configure it based on the context (AudioUnitSetProperty)
- Initialize the audio unit (AudioUnitInitialize)
Once the host is finished with an audio unit, it closes it:
- Dispose the audio unit (AudioComponentInstanceDispose)
Audio units can be used programmatically (for instance, mixers can be used to render audio for a
game, a generator to play audio files, etc), or they can be hosted in Digital Audio Workstation
(DAW) applications such as Logic, Garage Band, etc. In the DAW case, it is common for an audio
unit to provide a custom view to allow the user to interact with what can be complex DSP
operations that the audio unit performs. The view is retrieved from an audio unit through
AudioUnitGetProperty and then the host instantiates it (see <AudioToolbox/AUCocoaUIView.h>).
}
{ Pascal Translation: Gorazd Krosl <gorazd_1957@yahoo.ca>, October 2009 }
{ Pascal Translation Update: Jonas Maebe <jonas@freepascal.org>, October 2012 }
{ Pascal Translation Update: Jonas Maebe <jonas@freepascal.org>, July 2019 }
{
Modified for use with Free Pascal
Version 308
Please report any bugs to <gpc@microbizz.nl>
}
{$ifc not defined MACOSALLINCLUDE or not MACOSALLINCLUDE}
{$mode macpas}
{$modeswitch cblocks}
{$packenum 1}
{$macro on}
{$inline on}
{$calling mwpascal}
unit AUComponent;
interface
{$setc UNIVERSAL_INTERFACES_VERSION := $0400}
{$setc GAP_INTERFACES_VERSION := $0308}
{$ifc not defined USE_CFSTR_CONSTANT_MACROS}
{$setc USE_CFSTR_CONSTANT_MACROS := TRUE}
{$endc}
{$ifc defined CPUPOWERPC and defined CPUI386}
{$error Conflicting initial definitions for CPUPOWERPC and CPUI386}
{$endc}
{$ifc defined FPC_BIG_ENDIAN and defined FPC_LITTLE_ENDIAN}
{$error Conflicting initial definitions for FPC_BIG_ENDIAN and FPC_LITTLE_ENDIAN}
{$endc}
{$ifc not defined __ppc__ and defined CPUPOWERPC32}
{$setc __ppc__ := 1}
{$elsec}
{$setc __ppc__ := 0}
{$endc}
{$ifc not defined __ppc64__ and defined CPUPOWERPC64}
{$setc __ppc64__ := 1}
{$elsec}
{$setc __ppc64__ := 0}
{$endc}
{$ifc not defined __i386__ and defined CPUI386}
{$setc __i386__ := 1}
{$elsec}
{$setc __i386__ := 0}
{$endc}
{$ifc not defined __x86_64__ and defined CPUX86_64}
{$setc __x86_64__ := 1}
{$elsec}
{$setc __x86_64__ := 0}
{$endc}
{$ifc not defined __arm__ and defined CPUARM}
{$setc __arm__ := 1}
{$elsec}
{$setc __arm__ := 0}
{$endc}
{$ifc not defined __arm64__ and defined CPUAARCH64}
{$setc __arm64__ := 1}
{$elsec}
{$setc __arm64__ := 0}
{$endc}
{$ifc defined cpu64}
{$setc __LP64__ := 1}
{$elsec}
{$setc __LP64__ := 0}
{$endc}
{$ifc defined __ppc__ and __ppc__ and defined __i386__ and __i386__}
{$error Conflicting definitions for __ppc__ and __i386__}
{$endc}
{$ifc defined __ppc__ and __ppc__}
{$setc TARGET_CPU_PPC := TRUE}
{$setc TARGET_CPU_PPC64 := FALSE}
{$setc TARGET_CPU_X86 := FALSE}
{$setc TARGET_CPU_X86_64 := FALSE}
{$setc TARGET_CPU_ARM := FALSE}
{$setc TARGET_CPU_ARM64 := FALSE}
{$setc TARGET_OS_MAC := TRUE}
{$setc TARGET_OS_IPHONE := FALSE}
{$setc TARGET_IPHONE_SIMULATOR := FALSE}
{$setc TARGET_OS_EMBEDDED := FALSE}
{$elifc defined __ppc64__ and __ppc64__}
{$setc TARGET_CPU_PPC := FALSE}
{$setc TARGET_CPU_PPC64 := TRUE}
{$setc TARGET_CPU_X86 := FALSE}
{$setc TARGET_CPU_X86_64 := FALSE}
{$setc TARGET_CPU_ARM := FALSE}
{$setc TARGET_CPU_ARM64 := FALSE}
{$setc TARGET_OS_MAC := TRUE}
{$setc TARGET_OS_IPHONE := FALSE}
{$setc TARGET_IPHONE_SIMULATOR := FALSE}
{$setc TARGET_OS_EMBEDDED := FALSE}
{$elifc defined __i386__ and __i386__}
{$setc TARGET_CPU_PPC := FALSE}
{$setc TARGET_CPU_PPC64 := FALSE}
{$setc TARGET_CPU_X86 := TRUE}
{$setc TARGET_CPU_X86_64 := FALSE}
{$setc TARGET_CPU_ARM := FALSE}
{$setc TARGET_CPU_ARM64 := FALSE}
{$ifc defined(iphonesim)}
{$setc TARGET_OS_MAC := FALSE}
{$setc TARGET_OS_IPHONE := TRUE}
{$setc TARGET_IPHONE_SIMULATOR := TRUE}
{$elsec}
{$setc TARGET_OS_MAC := TRUE}
{$setc TARGET_OS_IPHONE := FALSE}
{$setc TARGET_IPHONE_SIMULATOR := FALSE}
{$endc}
{$setc TARGET_OS_EMBEDDED := FALSE}
{$elifc defined __x86_64__ and __x86_64__}
{$setc TARGET_CPU_PPC := FALSE}
{$setc TARGET_CPU_PPC64 := FALSE}
{$setc TARGET_CPU_X86 := FALSE}
{$setc TARGET_CPU_X86_64 := TRUE}
{$setc TARGET_CPU_ARM := FALSE}
{$setc TARGET_CPU_ARM64 := FALSE}
{$ifc defined(iphonesim)}
{$setc TARGET_OS_MAC := FALSE}
{$setc TARGET_OS_IPHONE := TRUE}
{$setc TARGET_IPHONE_SIMULATOR := TRUE}
{$elsec}
{$setc TARGET_OS_MAC := TRUE}
{$setc TARGET_OS_IPHONE := FALSE}
{$setc TARGET_IPHONE_SIMULATOR := FALSE}
{$endc}
{$setc TARGET_OS_EMBEDDED := FALSE}
{$elifc defined __arm__ and __arm__}
{$setc TARGET_CPU_PPC := FALSE}
{$setc TARGET_CPU_PPC64 := FALSE}
{$setc TARGET_CPU_X86 := FALSE}
{$setc TARGET_CPU_X86_64 := FALSE}
{$setc TARGET_CPU_ARM := TRUE}
{$setc TARGET_CPU_ARM64 := FALSE}
{ will require compiler define when/if other Apple devices with ARM cpus ship }
{$setc TARGET_OS_MAC := FALSE}
{$setc TARGET_OS_IPHONE := TRUE}
{$setc TARGET_IPHONE_SIMULATOR := FALSE}
{$setc TARGET_OS_EMBEDDED := TRUE}
{$elifc defined __arm64__ and __arm64__}
{$setc TARGET_CPU_PPC := FALSE}
{$setc TARGET_CPU_PPC64 := FALSE}
{$setc TARGET_CPU_X86 := FALSE}
{$setc TARGET_CPU_X86_64 := FALSE}
{$setc TARGET_CPU_ARM := FALSE}
{$setc TARGET_CPU_ARM64 := TRUE}
{ will require compiler define when/if other Apple devices with ARM cpus ship }
{$setc TARGET_OS_MAC := FALSE}
{$setc TARGET_OS_IPHONE := TRUE}
{$setc TARGET_IPHONE_SIMULATOR := FALSE}
{$setc TARGET_OS_EMBEDDED := TRUE}
{$elsec}
{$error __ppc__ nor __ppc64__ nor __i386__ nor __x86_64__ nor __arm__ nor __arm64__ is defined.}
{$endc}
{$ifc defined __LP64__ and __LP64__ }
{$setc TARGET_CPU_64 := TRUE}
{$elsec}
{$setc TARGET_CPU_64 := FALSE}
{$endc}
{$ifc defined FPC_BIG_ENDIAN}
{$setc TARGET_RT_BIG_ENDIAN := TRUE}
{$setc TARGET_RT_LITTLE_ENDIAN := FALSE}
{$elifc defined FPC_LITTLE_ENDIAN}
{$setc TARGET_RT_BIG_ENDIAN := FALSE}
{$setc TARGET_RT_LITTLE_ENDIAN := TRUE}
{$elsec}
{$error Neither FPC_BIG_ENDIAN nor FPC_LITTLE_ENDIAN are defined.}
{$endc}
{$setc ACCESSOR_CALLS_ARE_FUNCTIONS := TRUE}
{$setc CALL_NOT_IN_CARBON := FALSE}
{$setc OLDROUTINENAMES := FALSE}
{$setc OPAQUE_TOOLBOX_STRUCTS := TRUE}
{$setc OPAQUE_UPP_TYPES := TRUE}
{$setc OTCARBONAPPLICATION := TRUE}
{$setc OTKERNEL := FALSE}
{$setc PM_USE_SESSION_APIS := TRUE}
{$setc TARGET_API_MAC_CARBON := TRUE}
{$setc TARGET_API_MAC_OS8 := FALSE}
{$setc TARGET_API_MAC_OSX := TRUE}
{$setc TARGET_CARBON := TRUE}
{$setc TARGET_CPU_68K := FALSE}
{$setc TARGET_CPU_MIPS := FALSE}
{$setc TARGET_CPU_SPARC := FALSE}
{$setc TARGET_OS_UNIX := FALSE}
{$setc TARGET_OS_WIN32 := FALSE}
{$setc TARGET_RT_MAC_68881 := FALSE}
{$setc TARGET_RT_MAC_CFM := FALSE}
{$setc TARGET_RT_MAC_MACHO := TRUE}
{$setc TYPED_FUNCTION_POINTERS := TRUE}
{$setc TYPE_BOOL := FALSE}
{$setc TYPE_EXTENDED := FALSE}
{$setc TYPE_LONGLONG := TRUE}
uses MacTypes,AudioComponents,CFBase,CFArray,CFDate,CoreAudioTypes;
{$endc} {not MACOSALLINCLUDE}
{$ALIGN POWER}
//=====================================================================================================================
//#pragma mark Overview
//CF_ASSUME_NONNULL_BEGIN
//#define AU_SUPPORT_INTERAPP_AUDIO (TARGET_OS_IPHONE && !(0 && !TARGET_IPHONE_SIMULATOR && !TARGET_OS_EMBEDDED))
//================================================================================================
//#pragma mark -
//#pragma mark Audio Unit Types
{!
@typedef AudioUnit
@discussion An audio unit is of type AudioComponentInstance as defined in
AudioComponent.h
}
type
AudioUnit = AudioComponentInstance;
{!
@enum Audio Unit Types
@abstract different types of audio units
@discussion Audio units are classified into different types, where those types perform
different roles and functions.
There are some general categories of functionality that apply across different
types of audio units:
(1) Real-time usage
The audio unit will complete its operations in less time than is
represented by the render buffer. All audio units with the exception of
the OfflineEffect should meet this criteria
(2) Real-time I/O
Will request the same amount of audio input as it is being asked to
produce for output. Effects, Panners, Mixers and MusicDevices are required
to adhere to this restriction. FormatConverter's can with some constraints
be used in this situation (for instance, sample rate conversion, float-int),
but the host of the audio unit is responsible for insuring this.
(3) UI versus Programmatic usage
UI usage covers the case of using an audio unit in a Digital Audio
Workstation (DAW) with appropriate UI (for example a filter in Garage Band
or Logic). Effects, Panners, MusicDevices are all expected to be usable
within this context.
Programmatic usage is where an audio unit is used by a host app as part of
a general signal processing chain.
For instance, a mixer audio unit can be used to take several different
audio sources in a game and mix them together. Mixers, Output units are
for programmatic usage. FormatConverter and Generator types are generally
programmatic audio units, but if they can be used in a UI situation, they
specify a custom view. The app can then use that to decide that, with
appropriate constraints, the audio unit could be presented in a DAW type
application. For instance, the AUConverter audio unit from apple can do
sample rate conversion, etc, but has not general utility for a user in a
DAW app. Apple's Varispeed or AUTimePitch audio units can be used to change
the playback rate and pitch and so could be used to good affect by a user
in a DAW type environment, as well as just providing this general
functionality to any program.
@constant kAudioUnitType_Output
An output unit can be used as the head of an AUGraph. Apple provides a number
of output units that interface directly with an audio device
@constant kAudioUnitType_MusicDevice
Used to describe software musical instruments such as samplers and
synthesisers. They respond to MIDI and create notes, which are then controlled
through parameters or MIDI control messages. See <AudioToolbox/MusicDevice.h>
@constant kAudioUnitType_MusicEffect
Is an effect that is also able to respond directly to MIDI control messages,
typically through the mapping of these MIDI messages to different parameters
of the effect's DSP algorithm.
@constant kAudioUnitType_FormatConverter
A format converter is a general category for audio units that can change the
format (for instance, sample rate conversion) from an input to an output, as
well as other, non-I/O type manipulations (like a deferred render or varispeed
type of operation). As such, a format converter can ask for as much or as
little audio input to produce a given output. They are still expected to
complete their rendering within the time represented by the output buffer.
For format converters that have some utility as an "audio effect or processor",
it is quite common to provide an offline version of this audio unit as well.
For instance, Apple ships a format converter (for use in a "real-time" like
situation) and an offline version (for processing audio files) of the Time
Pitch and Varispeed audio units.
@constant kAudioUnitType_Effect
An audio unit that will process some x number of audio input samples to produce
x number of audio output samples. The common case for an effect is to have a
single input to a single output, though some effects take side-chain inputs as
well. Effects can be run in "offline" contexts (such as processing a file), but
they are expected to run in real-time. A delay unit or reverb is a good
example of this.
@constant kAudioUnitType_Mixer
An audio unit that takes some number of inputs, mixing them to provide 1 or
more audio outputs. A stere mixer (mono and stereo inputs to produce one
stereo output) is an example of this.
@constant kAudioUnitType_Panner
A panner is a specialised effect that will pan a single audio input to a single
output. Panner units are required to support a collection of standardised
parameters that specify the panning coordinates (aside from whatever custom
parameters the panner may provide). A surround panner is an example of this
@constant kAudioUnitType_Generator
A generator will have no audio input, but will just produce audio output. In
some ways it is similar to a MusicDevice, except that a generator provides no
MIDI input, or notion of "notes". A tone generator is a good example of this.
@constant kAudioUnitType_OfflineEffect
An offline effect is used to process data from a file and is also used to
publish a capability that cannot be run in real-time. For instance, the process
of normalisation requires seeing the entire audio input before the scalar to
apply in the normalisation process can be estimated. As such, offline effects
also have a notion of a priming stage that can be performed before the actual
rendering/processing phase is executed.
@constant kAudioUnitType_MIDIProcessor
Plugins of this type process midi input and produce midi output. They do not produce audio.
}
const
kAudioUnitType_Output = FourCharCode('auou');
kAudioUnitType_MusicDevice = FourCharCode('aumu');
kAudioUnitType_MusicEffect = FourCharCode('aumf');
kAudioUnitType_FormatConverter = FourCharCode('aufc');
kAudioUnitType_Effect = FourCharCode('aufx');
kAudioUnitType_Mixer = FourCharCode('aumx');
kAudioUnitType_Panner = FourCharCode('aupn');
kAudioUnitType_Generator = FourCharCode('augn');
kAudioUnitType_OfflineEffect = FourCharCode('auol');
kAudioUnitType_MIDIProcessor = FourCharCode('aumi');
{$ifc TARGET_OS_IPHONE and not TARGET_IPHONE_SIMULATOR}
{!
@enum Audio Unit Types (for inter-app audio)
@abstract types of inter-app audio units
@constant kAudioUnitType_RemoteEffect
@constant kAudioUnitType_RemoteGenerator
@constant kAudioUnitType_RemoteInstrument
@constant kAudioUnitType_RemoteMusicEffect
@discussion
These Audio Unit types are identical to the corresponding "non-remote" types (e.g.
kAudioUnitType_Effect, etc.), with the exception that they are the types that must be
used for AudioUnits defined by applications calling AudioOutputUnitPublish.
When Audio Components of these types are instantiated via AudioComponentInstanceNew,
a connection is made to the process which published the component via AudioOutputUnitPublish.
When using Audio Units which are instances of these components, one must take care to
initialize the unit only immediately before beginning a series of render operations, and
uninitialize it immediately afterwards, since while initialized, the Audio Unit's background
process is being kept awake and is consuming system resources.
When using AudioUnitGetProperty and AudioUnitSetProperty, only Apple-defined properties
are supported.
For kAudioUnitProperty_HostCallbacks, hosts can set this property on any remote audio unit.
This will cause the host callbacks to be called on each render cycle and the results
communicated to the remote AU's process. The owner of the published AU ("node") can *get*
this property on its output units, obtaining the structure of function pointers, and call
the host callbacks during the render cycle.
}
const
kAudioUnitType_RemoteEffect = FourCharCode('aurx');
kAudioUnitType_RemoteGenerator = FourCharCode('aurg');
kAudioUnitType_RemoteInstrument = FourCharCode('auri');
kAudioUnitType_RemoteMusicEffect = FourCharCode('aurm');
{$endc} {TARGET_OS_IPHONE and not TARGET_IPHONE_SIMULATOR}
//================================================================================================
//#pragma mark -
//#pragma mark Apple Audio Units
{!
@enum Apple audio unit manufacturer ID.
@discussion the unique ID used to identifier audio units provided by Apple, Inc.
}
const
kAudioUnitManufacturer_Apple = FourCharCode('appl');
{!
@enum Apple input/output audio unit sub types
@discussion These are the subtypes for the various input/output units that Apple ships. Input/output
units add an additional notion of Start and Stop.
see <AudioToolbox/AudioOutputUnit.h>
@constant kAudioUnitSubType_GenericOutput
A generic output unit provides the start/stop API, and provides the basic
services to convert Linear PCM formats.
@constant kAudioUnitSubType_VoiceProcessingIO
- Available on OS X and with iOS 3.0 or greater
This audio unit can do input as well as output. Bus 0 is used for the output
side, bus 1 is used to get audio input (thus, on the iPhone, it works in a
very similar way to the Remote I/O). This audio unit does signal processing on
the incoming audio (taking out any of the audio that is played from the device
at a given time from the incoming audio).
}
const
kAudioUnitSubType_GenericOutput = FourCharCode('genr');
kAudioUnitSubType_VoiceProcessingIO = FourCharCode('vpio');
{!
@enum Apple input/output audio unit sub types (OS X)
@constant kAudioUnitSubType_HALOutput
- desktop only
The audio unit that interfaces to any audio device. The user specifies which
audio device to track. The audio unit can do input from the device as well as
output to the device. Bus 0 is used for the output side, bus 1 is used
to get audio input from the device.
@constant kAudioUnitSubType_DefaultOutput
- desktop only
A specialisation of AUHAL that is used to track the user's selection of the
default device as set in the Sound Prefs
@constant kAudioUnitSubType_SystemOutput
- desktop only
A specialisation of AUHAL that is used to track the user's selection of the
device to use for sound effects, alerts
and other UI sounds.
}
const
{$ifc not TARGET_OS_IPHONE}
kAudioUnitSubType_HALOutput = FourCharCode('ahal');
kAudioUnitSubType_DefaultOutput = FourCharCode('def ');
kAudioUnitSubType_SystemOutput = FourCharCode('sys ');
{$endc} { not TARGET_OS_IPHONE }
{!
@enum Apple music instrument audio unit sub types
@constant kAudioUnitSubType_DLSSynth
- desktop only
A multi-timbral music device that can use sample banks in either DLS or
SoundFont formats. It fully supports GM-MIDI and the basic extensions of
GS-MIDI.
@constant kAudioUnitSubType_Sampler
A mono-timbral music device which is a sampler synthesizer and supports full
interactive editing of all state.
@constant kAudioUnitSubType_MIDISynth
A fully GM-compatible multi-timbral music device which is a sampler synthesizer.
It can load instruments from sample banks in either DLS or SoundFont formats.
}
//#if !TARGET_OS_IPHONE
{$ifc not TARGET_OS_IPHONE}
const
kAudioUnitSubType_DLSSynth = FourCharCode('dls ');
//#endif
{$endc} { not TARGET_OS_IPHONE }
kAudioUnitSubType_Sampler = FourCharCode('samp');
kAudioUnitSubType_MIDISynth = FourCharCode('msyn');
{!
@enum Apple converter audio unit sub types
@discussion These are the subtypes for the various converter units that apple ships.
Except for AUConverter, which is available on both desktop and iPhone, these
audio units are only available on the desktop.
@constant kAudioUnitSubType_AUConverter
An audio unit that uses an AudioConverter to do Linear PCM conversions (sample
rate, bit depth, interleaving).
@constant kAudioUnitSubType_Varispeed
An audio unit that can be used to control playback rate (as the rate is faster,
the pitch is higher). It provides a generic view, so can be used in both a UI
and programmatic context. It also comes in an Offline version so can be used
to process audio files.
@constant kAudioUnitSubType_DeferredRenderer
An audio unit that is used to get its input from a separate thread than the
thread that its render method is called. It thus allows an application to
introduce multiple threads into a rendering graph. There is a buffer sized
delay introduced between the input and output
@constant kAudioUnitSubType_Splitter
An audio unit that provides 2 output buses and 1 input bus. The audio unit
splits (duplicates) the input signal to the two output buses
@constant kAudioUnitSubType_MultiSplitter
An audio unit that sends its input bus to any number of output buses.
Every output bus gets all channels of the input bus.
This unit's implementation is lighter weight than kAudioUnitSubType_Splitter
even for two output buses, and is recommended over kAudioUnitSubType_Splitter.
@constant kAudioUnitSubType_Merger
An audio unit that provides 2 input buses and 2 output bus. The audio unit
merges the two inputs to the single output
@constant kAudioUnitSubType_NewTimePitch
An audio unit that provides good quality time stretching and pitch shifting
while still being very fast.
@constant kAudioUnitSubType_AUiPodTimeOther
An audio unit that provides time domain time stretching.
}
const
kAudioUnitSubType_AUConverter = FourCharCode('conv');
kAudioUnitSubType_Varispeed = FourCharCode('vari');
kAudioUnitSubType_DeferredRenderer = FourCharCode('defr');
kAudioUnitSubType_Splitter = FourCharCode('splt');
kAudioUnitSubType_MultiSplitter = FourCharCode('mspl');
kAudioUnitSubType_Merger = FourCharCode('merg');
kAudioUnitSubType_NewTimePitch = FourCharCode('nutp');
kAudioUnitSubType_AUiPodTimeOther = FourCharCode('ipto');
kAudioUnitSubType_RoundTripAAC = FourCharCode('raac');
{!
@enum Apple converter audio unit sub types (OS X only)
@constant kAudioUnitSubType_TimePitch
An audio unit that can be used to have independent control of both playback
rate and pitch. It provides a generic view, so can be used in both a UI and
programmatic context. It also comes in an Offline version so can be used to
process audio files.
}
{$ifc TARGET_OS_MAC}
const
kAudioUnitSubType_TimePitch = FourCharCode('tmpt');
{$endc} {TARGET_OS_MAC}
{!
@enum Apple effect audio unit sub types
@discussion These are the subtypes for the various effect units that apple ships
@constant kAudioUnitSubType_Delay
- desktop only
A delay audio unit
@constant kAudioUnitSubType_LowPassFilter
A filter that passes frequencies below a specified cut-off frequency
@constant kAudioUnitSubType_HighPassFilter
A filter that passes frequencies above a specified cut-off frequency
@constant kAudioUnitSubType_BandPassFilter
A filter that passes frequencies between a low and high cut-off frequency.
@constant kAudioUnitSubType_HighShelfFilter
A filter that can be used to implement a "treble" control
@constant kAudioUnitSubType_LowShelfFilter
A filter that can be used to implement a "bass" control
@constant kAudioUnitSubType_ParametricEQ
A parametric EQ filter
@constant kAudioUnitSubType_PeakLimiter
A peak limiter
@constant kAudioUnitSubType_DynamicsProcessor
A dynamics compressor/expander
@constant kAudioUnitSubType_SampleDelay
- desktop only
A delay that is used to delay the input a specified number of samples until
the output
@constant kAudioUnitSubType_Distortion
A distortion audio unit
@constant kAudioUnitSubType_NBandEQ
A generalized N-band graphic EQ with specifiable filter types per-band
}
const
kAudioUnitSubType_PeakLimiter = FourCharCode('lmtr');
kAudioUnitSubType_DynamicsProcessor = FourCharCode('dcmp');
kAudioUnitSubType_LowPassFilter = FourCharCode('lpas');
kAudioUnitSubType_HighPassFilter = FourCharCode('hpas');
kAudioUnitSubType_BandPassFilter = FourCharCode('bpas');
kAudioUnitSubType_HighShelfFilter = FourCharCode('hshf');
kAudioUnitSubType_LowShelfFilter = FourCharCode('lshf');
kAudioUnitSubType_ParametricEQ = FourCharCode('pmeq');
kAudioUnitSubType_Distortion = FourCharCode('dist');
{$ifc TARGET_OS_MAC}
kAudioUnitSubType_Delay = FourCharCode('dely');
kAudioUnitSubType_SampleDelay = FourCharCode('sdly');
{$endc} {TARGET_OS_MAC}
kAudioUnitSubType_NBandEQ = FourCharCode('nbeq');
{!
@enum Apple effect audio unit sub types (OS X only)
@constant kAudioUnitSubType_GraphicEQ
A 10 or 31 band Graphic EQ
@constant kAudioUnitSubType_MultiBandCompressor
A 4 band compressor/expander
@constant kAudioUnitSubType_MatrixReverb
A reverb that can be used to simulate various and different spaces
@constant kAudioUnitSubType_Pitch
An audio unit used to change the pitch
@constant kAudioUnitSubType_AUFilter
A filter unit that combines 5 different filters (low, 3 mids, high)
@constant kAudioUnitSubType_NetSend
An audio unit that is used in conjunction with _NetReceive to send audio
across the network (or between different applications)
@constant kAudioUnitSubType_RogerBeep
An audio unit that can be used to emit a short tone in gaps between speech
- similar to the tones used in a walkie-talkie
}
const
kAudioUnitSubType_GraphicEQ = FourCharCode('greq');
kAudioUnitSubType_MultiBandCompressor = FourCharCode('mcmp');
kAudioUnitSubType_MatrixReverb = FourCharCode('mrev');
kAudioUnitSubType_Pitch = FourCharCode('tmpt');
kAudioUnitSubType_AUFilter = FourCharCode('filt');
kAudioUnitSubType_NetSend = FourCharCode('nsnd');
kAudioUnitSubType_RogerBeep = FourCharCode('rogr');
{!
@enum Apple mixer audio unit sub types
@discussion These are the subtypes for the various mixer units that apple ships
@constant kAudioUnitSubType_MultiChannelMixer
Can have any number of inputs, with any number of channels on any input to one
output bus with any number of channels.
@constant kAudioUnitSubType_SpatialMixer
Inputs that are mono will be panned around using 3D coordinates and parameters.
Stereo inputs are passed directly through to the output.
A single output is presented with 2, 4, 5, 6, 7 or 8 channels.
There is also a built in reverb.
@constant kAudioUnitSubType_MatrixMixer
Any number of input and output buses with any number of channels on any bus.
The mix is presented as a matrix of channels that can be controlled through
input volume per channel, "cross-point" volume (a given input channel to a
given output channel), output volume per channel and a global volume across
the whole matrix
}
const
kAudioUnitSubType_MultiChannelMixer = FourCharCode('mcmx');
kAudioUnitSubType_MatrixMixer = FourCharCode('mxmx');
kAudioUnitSubType_SpatialMixer = FourCharCode('3dem');
{!
@enum Apple mixer audio unit sub types (OS X only)
@constant kAudioUnitSubType_StereoMixer
Inputs can be mono or stereo. Single stereo output
@constant kAudioUnitSubType_3DMixer
(deprecated, use kAudioUnitSubType_SpatialMixer instead)
Inputs can be mono, in which case they can be panned around using 3D
coordinates and parameters.
Stereo inputs are passed directly through to the output.
4 channel "ambisonic" inputs will be rendered to the output configuration
A single output of 2, 4, 5, 6, 7 or 8 channels.
}
const
kAudioUnitSubType_StereoMixer = FourCharCode('smxr');
kAudioUnitSubType_3DMixer = FourCharCode('3dmx'); (* API_DEPRECATED("no longer supported", macos(10.3, 10.10)) API_UNAVAILABLE(ios, watchos, tvos)
*)
// use kAudioUnitSubType_SpatialMixer instead
{!
@enum Apple panner audio unit sub types
@discussion These are the subtypes for the various panner units that apple ships
@constant kAudioUnitSubType_SphericalHeadPanner
- desktop only
A panner unit that uses the spherical head model to pan to a stereo output
@constant kAudioUnitSubType_VectorPanner
- desktop only
A panner unit that uses a moving pan between the two closes, adjacent channels
in a 3D space to a
surround output
@constant kAudioUnitSubType_SoundFieldPanner
- desktop only
A panner unit that uses a sound-field notion to pan to a surround output
@constant kAudioUnitSubType_HRTFPanner
- desktop only
A panner unit that uses a generic "HRTF" model to pan to a stereo output
}
{$if TARGET_OS_MAC}
const
kAudioUnitSubType_SphericalHeadPanner = FourCharCode('sphr');
kAudioUnitSubType_VectorPanner = FourCharCode('vbas');
kAudioUnitSubType_SoundFieldPanner = FourCharCode('ambi');
kAudioUnitSubType_HRTFPanner = FourCharCode('hrtf');
{$endc} { TARGET_OS_MAC }
{!
@enum Apple generator audio unit sub types
@discussion These are the subtypes for the various generator units that apple ships
@constant kAudioUnitSubType_ScheduledSoundPlayer
A generator unit that can be used to schedule slices of audio to be played at
a specified time. The audio is scheduled using the time stamps for the render
operation, and can be scheduled from any thread.
@constant kAudioUnitSubType_AudioFilePlayer
A generator unit that is used to play a file. It presents a custom UI so can
be used in a UI context as well
@constant kAudioUnitSubType_NetReceive
- desktop only
A generator unit that is paired with _NetSend to receive the audio that unit
sends. It presents a custom UI so can be used in a UI context as well
}
const
{$ifc TARGET_OS_MAC}
kAudioUnitSubType_NetReceive = FourCharCode('nrcv');
{$endc} { TARGET_OS_MAC }
kAudioUnitSubType_ScheduledSoundPlayer = FourCharCode('sspl');
kAudioUnitSubType_AudioFilePlayer = FourCharCode('afpl');
//=====================================================================================================================
//#pragma mark -
//#pragma mark Audio Unit Constants and typedefs
{!
@enum AudioUnitRenderActionFlags
@discussion These flags can be set in a callback from an audio unit during an audio unit
render operation from either the RenderNotify Proc or the render input
callback.
@constant kAudioUnitRenderAction_PreRender
Called on a render notification Proc - which is called either before or after
the render operation of the audio unit. If this flag is set, the proc is being
called before the render operation is performed.
@constant kAudioUnitRenderAction_PostRender
Called on a render notification Proc - which is called either before or after
the render operation of the audio unit. If this flag is set, the proc is being
called after the render operation is completed.
@constant kAudioUnitRenderAction_OutputIsSilence
This flag can be set in a render input callback (or in the audio unit's render
operation itself) and is used to indicate that the render buffer contains only
silence. It can then be used by the caller as a hint to whether the buffer
needs to be processed or not.
@constant kAudioOfflineUnitRenderAction_Preflight
This is used with offline audio units (of type 'auol'). It is used when an
offline unit is being preflighted, which is performed prior to the actual
offline rendering actions are performed. It is used for those cases where the
offline process needs it (for example, with an offline unit that normalises an
audio file, it needs to see all of the audio data first before it can perform
its normalisation)
@constant kAudioOfflineUnitRenderAction_Render
Once an offline unit has been successfully preflighted, it is then put into
its render mode. So this flag is set to indicate to the audio unit that it is
now in that state and that it should perform its processing on the input data.
@constant kAudioOfflineUnitRenderAction_Complete
This flag is set when an offline unit has completed either its preflight or
performed render operations
@constant kAudioUnitRenderAction_PostRenderError
If this flag is set on the post-render call an error was returned by the
AUs render operation. In this case, the error can be retrieved through the
lastRenderError property and the audio data in ioData handed to the post-render
notification will be invalid.
@constant kAudioUnitRenderAction_DoNotCheckRenderArgs
If this flag is set, then checks that are done on the arguments provided to render
are not performed. This can be useful to use to save computation time in
situations where you are sure you are providing the correct arguments
and structures to the various render calls
}
type
AudioUnitRenderActionFlags = UInt32;
AudioUnitRenderActionFlagsPtr = ^AudioUnitRenderActionFlags;
const
kAudioUnitRenderAction_PreRender = 1 shl 2;
kAudioUnitRenderAction_PostRender = 1 shl 3;
kAudioUnitRenderAction_OutputIsSilence = 1 shl 4;
kAudioOfflineUnitRenderAction_Preflight = 1 shl 5;
kAudioOfflineUnitRenderAction_Render = 1 shl 6;
kAudioOfflineUnitRenderAction_Complete = 1 shl 7;
kAudioUnitRenderAction_PostRenderError = 1 shl 8;
kAudioUnitRenderAction_DoNotCheckRenderArgs = 1 shl 9;
{!
@enum Audio unit errors
@discussion These are the various errors that can be returned by AudioUnit... API calls
@constant kAudioUnitErr_InvalidProperty
The property is not supported
@constant kAudioUnitErr_InvalidParameter
The parameter is not supported
@constant kAudioUnitErr_InvalidElement
The specified element is not valid
@constant kAudioUnitErr_NoConnection
There is no connection (generally an audio unit is asked to render but it has
not input from which to gather data)
@constant kAudioUnitErr_FailedInitialization
The audio unit is unable to be initialized
@constant kAudioUnitErr_TooManyFramesToProcess
When an audio unit is initialized it has a value which specifies the max
number of frames it will be asked to render at any given time. If an audio
unit is asked to render more than this, this error is returned.
@constant kAudioUnitErr_InvalidFile
If an audio unit uses external files as a data source, this error is returned
if a file is invalid (Apple's DLS synth returns this error)
@constant kAudioUnitErr_UnknownFileType
If an audio unit uses external files as a data source, this error is returned
if a file is invalid (Apple's DLS synth returns this error)
@constant kAudioUnitErr_FileNotSpecified
If an audio unit uses external files as a data source, this error is returned
if a file hasn't been set on it
(Apple's DLS synth returns this error)
@constant kAudioUnitErr_FormatNotSupported
Returned if an input or output format is not supported
@constant kAudioUnitErr_Uninitialized
Returned if an operation requires an audio unit to be initialized and it is
not.
@constant kAudioUnitErr_InvalidScope
The specified scope is invalid
@constant kAudioUnitErr_PropertyNotWritable
The property cannot be written
@constant kAudioUnitErr_CannotDoInCurrentContext
Returned when an audio unit is in a state where it can't perform the requested
action now - but it could later. Its usually used to guard a render operation
when a reconfiguration of its internal state is being performed.
@constant kAudioUnitErr_InvalidPropertyValue
The property is valid, but the value of the property being provided is not
@constant kAudioUnitErr_PropertyNotInUse
Returned when a property is valid, but it hasn't been set to a valid value at
this time.
@constant kAudioUnitErr_Initialized
Indicates the operation cannot be performed because the audio unit is
initialized.
@constant kAudioUnitErr_InvalidOfflineRender
Used to indicate that the offline render operation is invalid. For instance,
when the audio unit needs to be pre-flighted,
but it hasn't been.
@constant kAudioUnitErr_Unauthorized
Returned by either Open or Initialize, this error is used to indicate that the
audio unit is not authorised, that it cannot be used. A host can then present
a UI to notify the user the audio unit is not able to be used in its current
state.
@constant kAudioUnitErr_MIDIOutputBufferFull
Returned during the render call, if the audio unit produces more MIDI output,
than the default allocated buffer. The audio unit can provide a size hint, in
case it needs a larger buffer. See the documentation for AUAudioUnit's
MIDIOutputBufferSizeHint property.
@constant kAudioComponentErr_InstanceInvalidated
The component instance's implementation is not available, most likely because the process
that published it is no longer running.
@constant kAudioUnitErr_RenderTimeout
The audio unit did not satisfy the render request in time.
@constant kAudioUnitErr_ExtensionNotFound
The specified identifier did not match any Audio Unit Extensions.
@constant kAudioUnitErr_InvalidParameterValue
The parameter value is not supported, e.g. the value specified is NaN or
infinite.
}
const
kAudioUnitErr_InvalidProperty = -10879;
kAudioUnitErr_InvalidParameter = -10878;
kAudioUnitErr_InvalidElement = -10877;
kAudioUnitErr_NoConnection = -10876;
kAudioUnitErr_FailedInitialization = -10875;
kAudioUnitErr_TooManyFramesToProcess = -10874;
kAudioUnitErr_InvalidFile = -10871;
kAudioUnitErr_UnknownFileType = -10870;
kAudioUnitErr_FileNotSpecified = -10869;
kAudioUnitErr_FormatNotSupported = -10868;
kAudioUnitErr_Uninitialized = -10867;
kAudioUnitErr_InvalidScope = -10866;
kAudioUnitErr_PropertyNotWritable = -10865;
kAudioUnitErr_CannotDoInCurrentContext = -10863;
kAudioUnitErr_InvalidPropertyValue = -10851;
kAudioUnitErr_PropertyNotInUse = -10850;
kAudioUnitErr_Initialized = -10849;
kAudioUnitErr_InvalidOfflineRender = -10848;
kAudioUnitErr_Unauthorized = -10847;
kAudioUnitErr_MIDIOutputBufferFull = -66753;
kAudioComponentErr_InstanceTimedOut = -66754;
kAudioComponentErr_InstanceInvalidated = -66749;
kAudioUnitErr_RenderTimeout = -66745;
kAudioUnitErr_ExtensionNotFound = -66744;
kAudioUnitErr_InvalidParameterValue = -66743;
{$ifc TARGET_OS_IPHONE and not TARGET_IPHONE_SIMULATOR}
{!
@enum AudioComponent errors for inter-app audio
@constant kAudioComponentErr_DuplicateDescription
a non-unique component description was provided to AudioOutputUnitPublish
@constant kAudioComponentErr_UnsupportedType
an unsupported component type was provided to AudioOutputUnitPublish
@constant kAudioComponentErr_TooManyInstances
components published via AudioOutputUnitPublish may only have one instance
@constant kAudioComponentErr_NotPermitted
app needs "inter-app-audio" entitlement or host app needs "audio" in its UIBackgroundModes.
Or app is trying to register a component not declared in its Info.plist.
@constant kAudioComponentErr_InitializationTimedOut
host did not render in a timely manner; must uninitialize and reinitialize.
@constant kAudioComponentErr_InvalidFormat
inter-app AU element formats must have sample rates matching the hardware.
}
const
kAudioComponentErr_DuplicateDescription = -66752;
kAudioComponentErr_UnsupportedType = -66751;
kAudioComponentErr_TooManyInstances = -66750;
kAudioComponentErr_NotPermitted = -66748;
kAudioComponentErr_InitializationTimedOut = -66747;
kAudioComponentErr_InvalidFormat = -66746;
{$endc} {TARGET_OS_IPHONE and not TARGET_IPHONE_SIMULATOR}
{!
@typedef AudioUnitPropertyID
@discussion Type used for audio unit properties.
Properties are used to describe the state of an audio unit (for instance,
the input or output audio format)
}
type
AudioUnitPropertyID = UInt32;
{!
@typedef AudioUnitScope
@discussion Type used for audio unit scopes. Apple reserves the 0 < 1024 range for
audio unit scope identifiers.
Scopes are used to delineate a major attribute of an audio unit
(for instance, global, input, output)
}
type
AudioUnitScope = UInt32;
{!
@typedef AudioUnitElement
@discussion Type used for audio unit elements.
Scopes can have one or more member, and a member of a scope is
addressed / described by its element
For instance, input bus 1 is input scope, element 1
}
type
AudioUnitElement = UInt32;
{!
@typedef AudioUnitParameterID
@discussion Type used for audio unit parameters.
Parameters are typically used to control and set render state
(for instance, filter cut-off frequency)
}
type
AudioUnitParameterID = UInt32;
{!
@typedef AudioUnitParameterValue
@discussion Type used for audio unit parameter values.
The value of a given parameter is specified using this type
(typically a Float32)
}
type
AudioUnitParameterValue = Float32;
AudioUnitParameterValuePtr = ^AudioUnitParameterValue;
{!
@enum AUParameterEventType
@discussion The type of a parameter event (see AudioUnitScheduleParameter)
@constant kParameterEvent_Immediate
The parameter event describes an immediate change to the parameter value to
the new value
@constant kParameterEvent_Ramped
The parameter event describes a change to the parameter value that should
be applied over the specified period of time
}
type
AUParameterEventType = UInt32;
AUParameterEventTypePtr = ^AUParameterEventType;
const
kParameterEvent_Immediate = 1;
kParameterEvent_Ramped = 2;
{!
@struct AudioUnitParameterEvent
@discussion A parameter event describes a change to a parameter's value, where the type of
the event describes how that change is to be applied (see AUParameterEventType).
A parameter is uniquely defined through the triplet of scope, element and
parameterID.
See AudioUnitScheduleParameters
@field scope
The scope for the parameter
@field element
The element for the parameter
@field parameter
The parameterID for the parameter
@field eventType
The event type. This field further defines how the union described by
eventValues is to be interpreted.
@field eventValues
If the parameter event type is _Immediate, then the immediate struct of this
union should be used.
If the parameter event type is _Ramped, then the ramp struct of this union
should be used.
}
type
AudioUnitParameterEventRampRec = record
startBufferOffset: SInt32;
durationInFrames: UInt32;
startValue: AudioUnitParameterValue;
endValue: AudioUnitParameterValue;
end;
AudioUnitParameterEventImmediateRec = record
bufferOffset: UInt32;
value: AudioUnitParameterValue;
end;
AudioUnitParameterEventEventValues = record
case UInt8 of
0: (ramp : AudioUnitParameterEventRampRec);
1: (immediate : AudioUnitParameterEventImmediateRec);
end;
AudioUnitParameterEvent = record
scope: AudioUnitScope;
element: AudioUnitElement;
parameter: AudioUnitParameterID;
eventType: AUParameterEventType;
eventValues: AudioUnitParameterEventEventValues;
end;
AudioUnitParameterEventPtr = ^AudioUnitParameterEvent;
{!
@struct AudioUnitParameter
@discussion An audio unit parameter is defined by the triplet of audio unit scope, element
and parameterID. This struct is used with the functions in AudioUnitUtilities.h
to deal with audio unit parameters, but is included in this header file for
completeness.
@field mAudioUnit
The audio unit instance to which the specified parameter applies.
@field mParameterID
The parameterID for the parameter
@field mScope
The scope for the parameter
@field mElement
The element for the parameter
}
type
AudioUnitParameter = record
mAudioUnit: AudioUnit;
mParameterID: AudioUnitParameterID;
mScope: AudioUnitScope;
mElement: AudioUnitElement;
end;
AudioUnitParameterPtr = ^AudioUnitParameter;
{!
@struct AudioUnitProperty
@discussion An audio unit property is defined by the triplet of audio unit scope, element
and propertyID. This struct is used with the functions in AudioUnitUtilities.h
to deal with audio unit properties, but is included in this header file for
completeness.
@field mAudioUnit
The audio unit instance which the specified property applies too
@field mPropertyID
The propertyID for the property
@field mScope
The scope for the property
@field mElement
The element for the property
}
type
AudioUnitProperty = record
mAudioUnit: AudioUnit;
mPropertyID: AudioUnitPropertyID;
mScope: AudioUnitScope;
mElement: AudioUnitElement;
end;
AudioUnitPropertyPtr = ^AudioUnitProperty;
{!
@typedef AURenderCallback
@discussion This is the prototype for a function callback Proc that is used both with the
AudioUnit render notification API and the render input callback. See
kAudioUnitProperty_SetRenderCallback property or AudioUnitAddRenderNotify.
This callback is part of the process of a call to AudioUnitRender. As a
notification it is called either before or after the audio unit's render
operations. As a render input callback, it is called to provide input data for
the particular input bus the callback is attached too.
@param inRefCon
The client data that is provided either with the AURenderCallbackStruct or as
specified with the Add API call
@param ioActionFlags
Flags used to describe more about the context of this call (pre or post in the
notify case for instance)
@param inTimeStamp
The times stamp associated with this call of audio unit render
@param inBusNumber
The bus number associated with this call of audio unit render
@param inNumberFrames
The number of sample frames that will be represented in the audio data in the
provided ioData parameter
@param ioData
The AudioBufferList that will be used to contain the rendered or provided
audio data. These buffers will be aligned to 16 byte boundaries (which is
normally what malloc will return). Can be null in the notification that
input is available.
}
type
AURenderCallback = function( inRefCon: UnivPtr; var ioActionFlags: AudioUnitRenderActionFlags; const (*var*) inTimeStamp: AudioTimeStamp; inBusNumber: UInt32; inNumberFrames: UInt32; ioData: AudioBufferListPtr ): OSStatus;
{!
@typedef AudioUnitPropertyListenerProc
@discussion This is the prototype for a function callback Proc that is registered with an
audio unit to notify the caller of any changes to a value of an audio unit
property. See AudioUnitAddPropertyListener
@param inRefCon
The client data that is provided with the add property listener registration
@param inUnit
The audio unit upon which the specified property value has changed
@param inID
The property whose value has changed
@param inScope
The scope of the property whose value has changed
@param inElement
The element ID on the scope of the property whose value has changed
}
type
AudioUnitPropertyListenerProc = procedure( inRefCon: UnivPtr; inUnit: AudioUnit; inID: AudioUnitPropertyID; inScope: AudioUnitScope; inElement: AudioUnitElement );
{!
@typedef AUInputSamplesInOutputCallback
@discussion This is the prototype for a function callback Proc that is registered with an
audio unit to notify the caller of for the user of a varispeed or AUTimePitch
audio unit where it is not clear what input sample is represented in the
rendered output samples.
@param inRefCon
The client data that is provided with the add property listener registration
@param inOutputTimeStamp
The time stamp that corresponds to the first sample of audio data produced in
AudioUnitRender (its output data)
@param inInputSample
The sample number of the input that is represented in the first sample of that
output time stamp
@param inNumberInputSamples
The number of input samples that are represented in an output buffer
}
type
AUInputSamplesInOutputCallback = procedure( inRefCon: UnivPtr; const (*var*) inOutputTimeStamp: AudioTimeStamp; inInputSample: Float64; inNumberInputSamples: Float64 );
{!
@constant kAudioComponentRegistrationsChangedNotification
@abstract Notification generated when the set of available AudioComponents changes.
@discussion
Register for this notification name with [NSNotificationCenter defaultCenter] or
CFNotificationCenterGetLocalCenter(), using an object of NULL.
}
var kAudioComponentRegistrationsChangedNotification: CFStringRef; external name '_kAudioComponentRegistrationsChangedNotification'; (* attribute const *)
(* API_AVAILABLE(macos(10.11), ios(7.0), watchos(2.0), tvos(9.0)) *)
{!
@constant kAudioComponentInstanceInvalidationNotification
@abstract Notification generated when an audio unit extension process exits abnormally.
@discussion
Register for this notification name with [NSNotificationCenter defaultCenter] or
CFNotificationCenterGetLocalCenter(). The "object" refers to an AUAudioUnit instance to
be observed, or can be nil to observe all instances. The notification's userInfo
dictionary contains a key, "audioUnit", an NSValue whose pointerValue is the
AudioUnit or AudioComponentInstance which is wrapping the AUAudioUnit communicating with the
extension process. (This may be null if there is no such component instance.) For example:
[[NSNotificationCenter defaultCenter] addObserverForName:(NSString *)kAudioComponentInstanceInvalidationNotification object:nil queue:nil usingBlock:^(NSNotification *note) (
AUAudioUnit *auAudioUnit = (AUAudioUnit *)note.object;
NSValue *val = note.userInfo[@"audioUnit"];
AudioUnit audioUnit = (AudioUnit)val.pointerValue;
NSLog(@"Received kAudioComponentInstanceInvalidationNotification: auAudioUnit %@, audioUnit %p", auAudioUnit, audioUnit);
)];
}
var kAudioComponentInstanceInvalidationNotification: CFStringRef; external name '_kAudioComponentInstanceInvalidationNotification'; (* attribute const *)
(* API_AVAILABLE(macos(10.11), ios(9.0), watchos(2.0), tvos(9.0)) *)
//================================================================================================
//#pragma mark -
//#pragma mark Functions
{!
@function AudioUnitInitialize
@abstract initialize an audio unit
@discussion Upon success, the audio unit has been successfully initialized. This means
that the formats for input and output are valid and can be supported and it
has based its allocations on the max number of frames that it is able to
render at any given time. Once initialized, it is in a state where it can be
asked to render.
In common practice, major state of an audio unit (such as its I/O formats,
memory allocations) cannot be changed while an audio unit is initialized.
@param inUnit
The audio unit to initialize
@result noErr, or an error representing the reasons why the audio unit was not able
to be initialized successfully
}
function AudioUnitInitialize( inUnit: AudioUnit ): OSStatus; external name '_AudioUnitInitialize';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitUninitialize
@abstract uninitialize an audio unit
@discussion Once an audio unit has been initialized, to change its state in response to
some kind of environmental change, the audio unit should be uninitialized.
This will have the effect of the audio unit de-allocating its resources.
The caller can then reconfigure the audio unit to match the new environment
(for instance, the sample rate to process audio is different than it was) and
then re-initialize the audio unit when those changes have been applied.
@param inUnit
The audio unit to uninitialize
@result noErr, or an error representing the reasons why the audio unit was not able
to be initialized successfully. Typically this call won't return an error
unless the audio unit in question is no longer valid.
}
function AudioUnitUninitialize( inUnit: AudioUnit ): OSStatus; external name '_AudioUnitUninitialize';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitGetPropertyInfo
@abstract retrieves information about a specified property
@discussion The API can be used to retrieve both the size of the property, and whether it
is writable or not. In order to get a general answer on the capability of an
audio unit, this function should be called before the audio unit
is initialized (as some properties are writable when the audio unit is
initialized, and others not)
@param inUnit
the audio unit
@param inID
the property identifier
@param inScope
the scope of the property
@param inElement
the element of the scope
@param outDataSize
if not null, then will retrieve the maximum size for the property. if null,
then it is ignored
@param outWritable
if not null, then will retrieve whether the property can be written or not.
if null, then it is ignored
@result noErr, or various audio unit errors related to properties
}
function AudioUnitGetPropertyInfo( inUnit: AudioUnit; inID: AudioUnitPropertyID; inScope: AudioUnitScope; inElement: AudioUnitElement; outDataSize: UInt32Ptr; outWritable: BooleanPtr ): OSStatus; external name '_AudioUnitGetPropertyInfo';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitGetProperty
@abstract retrieves the value of a specified property
@discussion The API can is used to retrieve the value of the property. Property values for
audio units are always passed by reference
@param inUnit
the audio unit
@param inID
the property identifier
@param inScope
the scope of the property
@param inElement
the element of the scope
@param outData
used to retrieve the value of the property. It should point to memory at least
as large as the value described by ioDataSize
@param ioDataSize
on input contains the size of the data pointed to by outData, on output, the
size of the data that was returned.
@result noErr, or various audio unit errors related to properties
}
function AudioUnitGetProperty( inUnit: AudioUnit; inID: AudioUnitPropertyID; inScope: AudioUnitScope; inElement: AudioUnitElement; outData: UnivPtr; var ioDataSize: UInt32 ): OSStatus; external name '_AudioUnitGetProperty';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitSetProperty
@abstract sets the value of a specified property
@discussion The API can is used to set the value of the property. Property values for
audio units are always passed by reference
@param inUnit
the audio unit
@param inID
the property identifier
@param inScope
the scope of the property
@param inElement
the element of the scope
@param inData
if not null, then is the new value for the property that will be set. If null,
then inDataSize should be zero, and the call is then used to remove a
previously set value for a property. This removal is only valid for
some properties, as most properties will always have a default value if not
set.
@param inDataSize
the size of the data being provided in inData
@result noErr, or various audio unit errors related to properties
}
function AudioUnitSetProperty( inUnit: AudioUnit; inID: AudioUnitPropertyID; inScope: AudioUnitScope; inElement: AudioUnitElement; inData: {const} UnivPtr; inDataSize: UInt32 ): OSStatus; external name '_AudioUnitSetProperty';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitAddPropertyListener
@abstract registration call to receive notifications for when a property changes
@discussion When an audio unit property value changes, a notification callback can be
called by the audio unit to inform interested parties that this event has
occurred. The notification is defined by the tuple of inProc and
inProcUserData as paired to the specified property ID, so the previously
defined AudioUnitRemovePropertyListener is deprecated because it didn't allow
for the provision of the inProcUserData to remove a given listener (so,
you should use AudioUnitRemovePropertyListenerWithUserData).
@param inUnit
the audio unit
@param inID
the property identifier
@param inProc
the procedure to call when the property changes (on any scope or element)
@param inProcUserData
the user data to provide with the callback
@result noErr, or various audio unit errors related to properties
}
function AudioUnitAddPropertyListener( inUnit: AudioUnit; inID: AudioUnitPropertyID; inProc: AudioUnitPropertyListenerProc; inProcUserData: UnivPtr ): OSStatus; external name '_AudioUnitAddPropertyListener';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitRemovePropertyListenerWithUserData
@abstract remove a previously registered property listener
@discussion Removes a previously registered property listener as specified by the inProc
and inProcUser data as paired to the property identifier
@param inUnit
the audio unit
@param inID
the property identifier
@param inProc
the procedure previously registered
@param inProcUserData
the user data paired with the provided inProc
@result noErr, or various audio unit errors related to properties
}
function AudioUnitRemovePropertyListenerWithUserData( inUnit: AudioUnit; inID: AudioUnitPropertyID; inProc: AudioUnitPropertyListenerProc; inProcUserData: UnivPtr ): OSStatus; external name '_AudioUnitRemovePropertyListenerWithUserData';
(* API_AVAILABLE(macos(10.5), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitAddRenderNotify
@abstract a notification callback to call when an audio unit is asked to render
@discussion allows an application to register a callback with an audio unit for whenever
the audio unit is asked to render. The callback is called both before the
audio unit performs its render operations (the render flag's pre-render bit
is set) and after the audio unit has completed its render operations (the
render flag's post-render bit is set). On post-render, the audio buffer list
(ioData) will contain valid audio data that was rendered by the audio unit.
The inProc and inProcUserData are treated as a tuple entity, so when wanting
to remove one, both the inProc and its inProcUserData must be specified
@param inUnit
the audio unit
@param inProc
an AURenderCallback proc
@param inProcUserData
the user data that will be provided with the proc when it is called
@result noErr, or an audio unit error code
}
function AudioUnitAddRenderNotify( inUnit: AudioUnit; inProc: AURenderCallback; inProcUserData: UnivPtr ): OSStatus; external name '_AudioUnitAddRenderNotify';
(* API_AVAILABLE(macos(10.2), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitRemoveRenderNotify
@abstract remove a previously registered render notification proc
@param inUnit
the audio unit
@param inProc
an AURenderCallback proc
@param inProcUserData
the user data that was provided with the proc when it was previously
registered
@result noErr, or an audio unit error code
}
function AudioUnitRemoveRenderNotify( inUnit: AudioUnit; inProc: AURenderCallback; inProcUserData: UnivPtr ): OSStatus; external name '_AudioUnitRemoveRenderNotify';
(* API_AVAILABLE(macos(10.2), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitGetParameter
@abstract Get the value of a parameter
@discussion Get the value of a parameter as specified by its ID, scope and element.
@param inUnit
the audio unit
@param inID
the parameter ID
@param inScope
the scope for the parameter
@param inElement
the element on the scope for the parameter
@param outValue
Must be non-null, and upon success will contain the current value for the
specified parameter
@result noErr, or an audio unit error code (such as InvalidParameter)
}
function AudioUnitGetParameter( inUnit: AudioUnit; inID: AudioUnitParameterID; inScope: AudioUnitScope; inElement: AudioUnitElement; var outValue: AudioUnitParameterValue ): OSStatus; external name '_AudioUnitGetParameter';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitSetParameter
@abstract Set the value of a parameter
@discussion Set the value of a parameter as specified by its ID, scope and element.
Parameter IDs are consistent across all of the elements in a scope - so for a
mixer, the "input volume" parameter can be applied on any input, and the
particular input is specified by the elementID.
@param inUnit
the audio unit
@param inID
the parameter ID
@param inScope
the scope for the parameter
@param inElement
the element on the scope for the parameter
@param inValue
the new value for the parameter.
@param inBufferOffsetInFrames
generally should be set to zero - see AudioUnitScheduleParameters
@result noErr, or an audio unit error code (such as InvalidParameter)
}
function AudioUnitSetParameter( inUnit: AudioUnit; inID: AudioUnitParameterID; inScope: AudioUnitScope; inElement: AudioUnitElement; inValue: AudioUnitParameterValue; inBufferOffsetInFrames: UInt32 ): OSStatus; external name '_AudioUnitSetParameter';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitScheduleParameters
@abstract Schedule changes to the value of a parameter
@discussion This API is used to schedule intra-buffer changes to the value of a parameter
(immediate) or to ramp a parameter from a start value to an end value for a
specified number of samples (ramp)
The API allows for the scheduling of multiple parameter events with the one
call. All of the parameter events must apply to the current (and only apply to
the current) audio unit render call, so the events are scheduled as a part of
the pre-render notification callback.
When scheduling an immediate parameter event, the new value at the specified
sample buffer offset is provided
When scheduling a ramped parameter, the ramp is scheduled each audio unit
render for the duration of the ramp. Each schedule of the the new audio unit
render specifies the progress of the ramp.
Parameters that can have events scheduled to them will indicate this through
their parameter info struct
@param inUnit
the audio unit
@param inParameterEvent
a pointer to an array of parameter event structs
@param inNumParamEvents
the number of parameter event structs pointed to by inParameterEvent
@result noErr, or an audio unit error code (such as InvalidParameter)
}
function AudioUnitScheduleParameters( inUnit: AudioUnit; const inParameterEvent: AudioUnitParameterEventPtr; inNumParamEvents: UInt32 ): OSStatus; external name '_AudioUnitScheduleParameters';
(* API_AVAILABLE(macos(10.2), ios(2.0), watchos(2.0), tvos(9.0)) *)
{!
@function AudioUnitRender
@abstract the render operation where ioData will contain the results of the audio unit's
render operations
@discussion an audio unit will render the amount of audio data described by
inNumberOfFrames and the results of that render will be contained within
ioData. The caller should provide audio time stamps where at least the sample
time is valid and it is incrementing sequentially from its previous call
(so, the next time stamp will be the current time stamp + inNumberFrames)
If the sample time is not incrementing sequentially, the audio unit will infer
that there is some discontinuity with the timeline it is rendering for
The caller must provide a valid ioData AudioBufferList that matches the
expected topology for the current audio format for the given bus. The buffer
list can be of two variants:
(1) If the mData pointers are non-null then the audio unit will render its
output into those buffers. These buffers should be aligned to 16 byte
boundaries (which is normally what malloc will return).
(2) If the mData pointers are null, then the audio unit can provide pointers
to its own buffers. In this case the audio unit is required to keep those
buffers valid for the duration of the calling thread's I/O cycle
@param inUnit
the audio unit
@param ioActionFlags
any appropriate action flags for the render operation
@param inTimeStamp
the time stamp that applies to this particular render operation. when
rendering for multiple output buses the time stamp will generally be the same
for each output bus, so the audio unit is able to determine without doubt that
this the same render operation
@param inOutputBusNumber
the output bus to render for
@param inNumberFrames
the number of sample frames to render
@param ioData
the audio buffer list that the audio unit is to render into.
@result noErr, or an audio unit render error
}
function AudioUnitRender( inUnit: AudioUnit; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inOutputBusNumber: UInt32; inNumberFrames: UInt32; ioData: AudioBufferListPtr ): OSStatus; external name '_AudioUnitRender';
(* API_AVAILABLE(macos(10.2), ios(2.0), watchos(2.0), tvos(9.0)) *)
{$ifc TARGET_OS_MAC}
function AudioUnitProcess( inUnit: AudioUnit; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inNumberFrames: UInt32; var ioData: AudioBufferList ): OSStatus; external name '_AudioUnitProcess';
(* API_AVAILABLE(macos(10.7), ios(6.0), watchos(2.0), tvos(9.0)) *)
function AudioUnitProcessMultiple( inUnit: AudioUnit; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inNumberFrames: UInt32; inNumberInputBufferLists: UInt32; {const} inInputBufferLists: AudioBufferListPtrPtr; inNumberOutputBufferLists: UInt32; ioOutputBufferLists: AudioBufferListPtrPtr ): OSStatus; external name '_AudioUnitProcessMultiple';
(* API_AVAILABLE(macos(10.7), ios(6.0), watchos(2.0), tvos(9.0)) *)
{$endc} {TARGET_OS_MAC}
{!
@function AudioUnitReset
@abstract reset an audio unit's render state
@discussion This call will clear any render state of an audio unit. For instance, with a
delay or reverb type of audio unit reset will clear any of the delay lines
maintained within the audio unit. Typically, this call is made when an
audio unit was previously rendering, and was taken out of the render chain
(say, the track it is in was muted) and is being added back in (unmuted).
The host should reset the audio unit before adding it back so that it doesn't
produce audio from its delay lines that is no longer valid.
The call should only clear memory, it should NOT allocate or free memory
resources (this is done in the Initialize calls).
@param inUnit
the audio unit
@param inScope
the scope - typically this is set to GlobalScope
@param inElement
the element - typically this is set to 0
@result noErr, or an audio unit error
}
function AudioUnitReset( inUnit: AudioUnit; inScope: AudioUnitScope; inElement: AudioUnitElement ): OSStatus; external name '_AudioUnitReset';
(* API_AVAILABLE(macos(10.0), ios(2.0), watchos(2.0), tvos(9.0)) *)
{$ifc TARGET_OS_IPHONE and not TARGET_IPHONE_SIMULATOR}
{!
@function AudioOutputUnitPublish
@abstract Register an audio output unit as available to be used as an audio unit by
other applications.
@param inOutputUnit
The audio output unit to be published.
@param inDesc
The AudioComponentDescription under which to register the application.
@param inName
The application or component name.
@result An OSStatus result code.
@discussion
This allows a publishing application to register its audio (input/)output unit as being able
to be redirected and repurposed as an audio unit effect, generator, music device or music
effect by another host application.
}
function AudioOutputUnitPublish( const (*var*) inDesc: AudioComponentDescription; inName: CFStringRef; inVersion: UInt32; inOutputUnit: AudioUnit ): OSStatus; external name '_AudioOutputUnitPublish';
(* API_AVAILABLE(ios(7.0), watchos(2.0), tvos(9.0)) API_UNAVAILABLE(macos) *)
{!
@function AudioComponentGetLastActiveTime
@abstract Fetches the time at which the application publishing the component was last active.
@discussion
Inter-app audio hosts can use this to sort the list of available nodes by how recently
the user interacted with them.
@param inComponent
The AudioComponent being queried.
@result The CFAbsoluteTime at which the node was last active (0 if never).
}
function AudioComponentGetLastActiveTime( comp: AudioComponent ): CFAbsoluteTime; external name '_AudioComponentGetLastActiveTime';
(* API_AVAILABLE(ios(7.0), watchos(2.0), tvos(9.0)) API_UNAVAILABLE(macos) *)
{$endc} {TARGET_OS_IPHONE and not TARGET_IPHONE_SIMULATOR}
{$ifc TARGET_CPU_64 or TARGET_OS_IPHONE}
{!
@function AudioUnitExtensionSetComponentList
@abstract Allows the implementor of an audio unit extension to dynamically modify the
list of component registrations for the extension.
@param extensionIdentifier
The bundle ID of the audio unit extension.
@param audioComponentInfo
An array of dictionaries, one for each component, in the same format as
described in AudioComponent.h for the Info.plist key "AudioComponents".
@result An OSStatus result code.
@discussion
Note that the bundle ID of the process calling this API must prefix (or match)
the provided extension identifier.
}
function AudioUnitExtensionSetComponentList(extensionIdentifier: CFStringRef; audioComponentInfo: CFArrayRef {__nullable}): OSStatus; external name '_AudioUnitExtensionSetComponentList';
(* API_AVAILABLE(macos(10.13), ios(11.0)) __TVOS_PROHIBITED __WATCHOS_PROHIBITED *)
{!
@function AudioUnitExtensionCopyComponentList
@abstract Returns the component registrations for a given audio unit extension.
@param extensionIdentifier
The bundle ID of the audio unit extension.
@result An array of dictionaries, one for each component, in the same format as
described in AudioComponent.h for the Info.plist key "AudioComponents".
The caller should release this value when done with it.
}
function AudioUnitExtensionCopyComponentList(extensionIdentifier: CFStringRef): CFArrayRef {__nullable}; external name '_AudioUnitExtensionCopyComponentList';
(* API_AVAILABLE(macos(10.13), ios(11.0)) __TVOS_PROHIBITED __WATCHOS_PROHIBITED *)
{$endc} {TARGET_CPU_64 or TARGET_OS_IPHONE}
{!
@enum AudioUnitRange
@discussion the range of selectors that are used to dispatch through to the various audio
unit API
@constant kAudioUnitRange
@constant kAudioUnitInitializeSelect
@constant kAudioUnitUninitializeSelect
@constant kAudioUnitGetPropertyInfoSelect
@constant kAudioUnitGetPropertySelect
@constant kAudioUnitSetPropertySelect
@constant kAudioUnitAddPropertyListenerSelect
@constant kAudioUnitRemovePropertyListenerSelect
@constant kAudioUnitRemovePropertyListenerWithUserDataSelect
@constant kAudioUnitAddRenderNotifySelect
@constant kAudioUnitRemoveRenderNotifySelect
@constant kAudioUnitGetParameterSelect
@constant kAudioUnitSetParameterSelect
@constant kAudioUnitScheduleParametersSelect
@constant kAudioUnitRenderSelect
@constant kAudioUnitResetSelect
@constant kAudioUnitComplexRenderSelect
@constant kAudioUnitProcessSelect
@constant kAudioUnitProcessMultipleSelect
}
const
kAudioUnitRange = $0000; // range of selectors for audio units
kAudioUnitInitializeSelect = $0001;
kAudioUnitUninitializeSelect = $0002;
kAudioUnitGetPropertyInfoSelect = $0003;
kAudioUnitGetPropertySelect = $0004;
kAudioUnitSetPropertySelect = $0005;
kAudioUnitAddPropertyListenerSelect = $000A;
kAudioUnitRemovePropertyListenerSelect = $000B;
kAudioUnitRemovePropertyListenerWithUserDataSelect = $0012;
kAudioUnitAddRenderNotifySelect = $000F;
kAudioUnitRemoveRenderNotifySelect = $0010;
kAudioUnitGetParameterSelect = $0006;
kAudioUnitSetParameterSelect = $0007;
kAudioUnitScheduleParametersSelect = $0011;
kAudioUnitRenderSelect = $000E;
kAudioUnitResetSelect = $0009;
kAudioUnitComplexRenderSelect = $0013;
kAudioUnitProcessSelect = $0014;
kAudioUnitProcessMultipleSelect = $0015;
//================================================================================================
//#pragma mark -
//#pragma mark Dispatch function prototypes
type
AudioUnitInitializeProc = function( self: UnivPtr ): OSStatus;
type
AudioUnitUninitializeProc = function( self: UnivPtr ): OSStatus;
type
AudioUnitGetPropertyInfoProc = function( self: UnivPtr; prop: AudioUnitPropertyID; scope: AudioUnitScope; elem: AudioUnitElement; outDataSize: UInt32Ptr; outWritable: BooleanPtr ): OSStatus;
type
AudioUnitGetPropertyProc = function( self: UnivPtr; inID: AudioUnitPropertyID; inScope: AudioUnitScope; inElement: AudioUnitElement; outData: UnivPtr; var ioDataSize: UInt32 ): OSStatus;
type
AudioUnitSetPropertyProc = function( self: UnivPtr; inID: AudioUnitPropertyID; inScope: AudioUnitScope; inElement: AudioUnitElement; inData: {const} UnivPtr; inDataSize: UInt32 ): OSStatus;
type
AudioUnitAddPropertyListenerProc = function( self: UnivPtr; prop: AudioUnitPropertyID; proc: AudioUnitPropertyListenerProc; userData: UnivPtr ): OSStatus;
type
AudioUnitRemovePropertyListenerProc = function( self: UnivPtr; prop: AudioUnitPropertyID; proc: AudioUnitPropertyListenerProc ): OSStatus;
type
AudioUnitRemovePropertyListenerWithUserDataProc = function( self: UnivPtr; prop: AudioUnitPropertyID; proc: AudioUnitPropertyListenerProc; userData: UnivPtr ): OSStatus;
type
AudioUnitAddRenderNotifyProc = function( self: UnivPtr; proc: AURenderCallback; userData: UnivPtr ): OSStatus;
type
AudioUnitRemoveRenderNotifyProc = function( self: UnivPtr; proc: AURenderCallback; userData: UnivPtr ): OSStatus;
type
AudioUnitScheduleParametersProc = function( self: UnivPtr; const (*var*) events: AudioUnitParameterEvent; numEvents: UInt32 ): OSStatus;
type
AudioUnitResetProc = function( self: UnivPtr; inScope: AudioUnitScope; inElement: AudioUnitElement ): OSStatus;
type
AudioUnitComplexRenderProc = function( self: UnivPtr; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inOutputBusNumber: UInt32; inNumberOfPackets: UInt32; var outNumberOfPackets: UInt32; var outPacketDescriptions: AudioStreamPacketDescription; var ioData: AudioBufferList; outMetadata: UnivPtr; var outMetadataByteSize: UInt32 ): OSStatus;
type
AudioUnitProcessProc = function( self: UnivPtr; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inNumberFrames: UInt32; var ioData: AudioBufferList ): OSStatus;
type
AudioUnitProcessMultipleProc = function( self: UnivPtr; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inNumberFrames: UInt32; inNumberInputBufferLists: UInt32; {const} inInputBufferLists: AudioBufferListPtrPtr; inNumberOutputBufferLists: UInt32; ioOutputBufferLists: AudioBufferListPtrPtr ): OSStatus;
{!
@typedef AudioUnitGetParameterProc
@discussion This proc can be exported through the FastDispatch property or is used as the prototype for
an audio component dispatch for this selector.
The arguments are the same as are provided to the corresponding API call
@param inComponentStorage
For a component manager component, this is the component instance storage
pointer
@param inID
@param inScope
@param inElement
@param outValue
}
type
AudioUnitGetParameterProc = function( inComponentStorage: UnivPtr; inID: AudioUnitParameterID; inScope: AudioUnitScope; inElement: AudioUnitElement; var outValue: AudioUnitParameterValue ): OSStatus;
{!
@typedef AudioUnitSetParameterProc
@discussion This proc can be exported through the FastDispatch property or is used as the prototype for
an audio component dispatch for this selector.
The arguments are the same as are provided to the corresponding API call
@param inComponentStorage
For a component manager component, this is the component instance storage
pointer
@param inID
@param inScope
@param inElement
@param inValue
@param inBufferOffsetInFrames
}
type
AudioUnitSetParameterProc = function( inComponentStorage: UnivPtr; inID: AudioUnitParameterID; inScope: AudioUnitScope; inElement: AudioUnitElement; inValue: AudioUnitParameterValue; inBufferOffsetInFrames: UInt32 ): OSStatus;
{!
@typedef AudioUnitRenderProc
@discussion This proc can be exported through the FastDispatch property or is used as the prototype for
an audio component dispatch for this selector.
The arguments are the same as are provided to the corresponding API call
@param inComponentStorage
For a component manager component, this is the component instance storage
pointer
@param ioActionFlags
@param inTimeStamp
@param inOutputBusNumber
@param inNumberFrames
@param ioData
}
type
AudioUnitRenderProc = function( inComponentStorage: UnivPtr; ioActionFlags: AudioUnitRenderActionFlagsPtr; const (*var*) inTimeStamp: AudioTimeStamp; inOutputBusNumber: UInt32; inNumberFrames: UInt32; var ioData: AudioBufferList ): OSStatus;
//=====================================================================================================================
//#pragma mark -
//#pragma mark Deprecated
{!
@enum deprecated - Audio unit errors
@discussion These are the various errors that can be returned by AudioUnit... API calls
@constant kAudioUnitErr_IllegalInstrument
Apple's DLS synth returns this error if information about a particular
instrument patch is requested, but is not valid.
@constant kAudioUnitErr_InstrumentTypeNotFound
Apple's DLS synth returns this error if information about a particular
instrument patch is requested, but is not valid.
}
const
kAudioUnitErr_IllegalInstrument = -10873;
kAudioUnitErr_InstrumentTypeNotFound = -10872;
//#if !__LP64__ && !TARGET_OS_IPHONE
{$ifc not TARGET_CPU_64 and not TARGET_OS_IPHONE}
// this call is deprecated and replaced by AudioUnitRemovePropertyListenerWithUserData
// this allows apps to use the same function pointer more than once
// you provide the same function ptr and user data as provided when you add a property listener
function AudioUnitRemovePropertyListener( inUnit: AudioUnit; inID: AudioUnitPropertyID; inProc: AudioUnitPropertyListenerProc ): OSStatus; external name '_AudioUnitRemovePropertyListener';
(* __OSX_AVAILABLE_BUT_DEPRECATED(__MAC_10_0,__MAC_10_5, __IPHONE_NA, __IPHONE_NA) *)
//#endif
{$endc}
{$ifc not defined MACOSALLINCLUDE or not MACOSALLINCLUDE}
end.
{$endc} {not MACOSALLINCLUDE}