AudioContext

Overview

The AudioContext interface represents an audio-processing graph built from audio modules linked together, each represented by an AudioNode.

Location

 

createAnalyser    (declared in BaseAudioContext)

Creates an AnalyserNode, which can be used to expose audio time and frequency data and for example to create data visualisations.

 

method createAnalyser: AnalyserNode

 

AnalyserNode createAnalyser()

 

func createAnalyser() -> AnalyserNode

 

AnalyserNode createAnalyser()

 

Function createAnalyser() As AnalyserNode

createBiquadFilter    (declared in BaseAudioContext)

Creates a BiquadFilterNode, which represents a second order filter configurable as several different common filter types: high-pass, low-pass, band-pass, etc

 

method createBiquadFilter: BiquadFilterNode

 

BiquadFilterNode createBiquadFilter()

 

func createBiquadFilter() -> BiquadFilterNode

 

BiquadFilterNode createBiquadFilter()

 

Function createBiquadFilter() As BiquadFilterNode

createBufferSource    (declared in BaseAudioContext)

Creates an AudioBufferSourceNode, which can be used to play and manipulate audio data contained within an AudioBuffer object. AudioBuffers are created using AudioContext.createBuffer or returned by AudioContext.decodeAudioData when it successfully decodes an audio track.

 

method createBufferSource: AudioBufferSourceNode

 

AudioBufferSourceNode createBufferSource()

 

func createBufferSource() -> AudioBufferSourceNode

 

AudioBufferSourceNode createBufferSource()

 

Function createBufferSource() As AudioBufferSourceNode

createChannelMerger    (declared in BaseAudioContext)

Creates a ChannelMergerNode, which is used to combine channels from multiple audio streams into a single audio stream.

 

method createChannelMerger(parnumberOfInputs: dynamic): ChannelMergerNode

 

ChannelMergerNode createChannelMerger(dynamic parnumberOfInputs)

 

func createChannelMerger(_ parnumberOfInputs: dynamic) -> ChannelMergerNode

 

ChannelMergerNode createChannelMerger(dynamic parnumberOfInputs)

 

Function createChannelMerger(parnumberOfInputs As dynamic) As ChannelMergerNode

Parameters:

  • parnumberOfInputs:

createChannelSplitter    (declared in BaseAudioContext)

Creates a ChannelSplitterNode, which is used to access the individual channels of an audio stream and process them separately.

 

method createChannelSplitter(parnumberOfOutputs: dynamic): ChannelSplitterNode

 

ChannelSplitterNode createChannelSplitter(dynamic parnumberOfOutputs)

 

func createChannelSplitter(_ parnumberOfOutputs: dynamic) -> ChannelSplitterNode

 

ChannelSplitterNode createChannelSplitter(dynamic parnumberOfOutputs)

 

Function createChannelSplitter(parnumberOfOutputs As dynamic) As ChannelSplitterNode

Parameters:

  • parnumberOfOutputs:

createConstantSource    (declared in BaseAudioContext)

Creates a ConstantSourceNode object, which is an audio source that continuously outputs a monaural (one-channel) sound signal whose samples all have the same value.

 

method createConstantSource: ConstantSourceNode

 

ConstantSourceNode createConstantSource()

 

func createConstantSource() -> ConstantSourceNode

 

ConstantSourceNode createConstantSource()

 

Function createConstantSource() As ConstantSourceNode

createConvolver    (declared in BaseAudioContext)

Creates a ConvolverNode, which can be used to apply convolution effects to your audio graph, for example a reverberation effect.

 

method createConvolver: ConvolverNode

 

ConvolverNode createConvolver()

 

func createConvolver() -> ConvolverNode

 

ConvolverNode createConvolver()

 

Function createConvolver() As ConvolverNode

createDelay    (declared in BaseAudioContext)

Creates a DelayNode, which is used to delay the incoming audio signal by a certain amount. This node is also useful to create feedback loops in a Web Audio API graph.

 

method createDelay(parmaxDelayTime: dynamic): DelayNode

 

DelayNode createDelay(dynamic parmaxDelayTime)

 

func createDelay(_ parmaxDelayTime: dynamic) -> DelayNode

 

DelayNode createDelay(dynamic parmaxDelayTime)

 

Function createDelay(parmaxDelayTime As dynamic) As DelayNode

Parameters:

  • parmaxDelayTime:

createDynamicsCompressor    (declared in BaseAudioContext)

Creates a DynamicsCompressorNode, which can be used to apply acoustic compression to an audio signal.

 

method createDynamicsCompressor: DynamicsCompressorNode

 

DynamicsCompressorNode createDynamicsCompressor()

 

func createDynamicsCompressor() -> DynamicsCompressorNode

 

DynamicsCompressorNode createDynamicsCompressor()

 

Function createDynamicsCompressor() As DynamicsCompressorNode

createGain    (declared in BaseAudioContext)

Creates a GainNode, which can be used to control the overall volume of the audio graph.

 

method createGain: GainNode

 

GainNode createGain()

 

func createGain() -> GainNode

 

GainNode createGain()

 

Function createGain() As GainNode

createIIRFilter    (declared in BaseAudioContext)

Creates an IIRFilterNode, which represents a second order filter configurable as several different common filter types.

 

method createIIRFilter(parfeedforward: dynamic; parfeedback: dynamic): IIRFilterNode

 

IIRFilterNode createIIRFilter(dynamic parfeedforward, dynamic parfeedback)

 

func createIIRFilter(_ parfeedforward: dynamic, _ parfeedback: dynamic) -> IIRFilterNode

 

IIRFilterNode createIIRFilter(dynamic parfeedforward, dynamic parfeedback)

 

Function createIIRFilter(parfeedforward As dynamic, parfeedback As dynamic) As IIRFilterNode

Parameters:

  • parfeedforward:
  • parfeedback:

createMediaElementSource

Creates a MediaElementAudioSourceNode associated with an HTMLMediaElement. This can be used to play and manipulate audio from ) embeds a media player which supports video playback into the document. You can use <video> for audio content as well, but the <audio> element may provide a more appropriate user experience."><video> or element is used to embed sound content in documents. It may contain one or more audio sources, represented using the src attribute or the <source> element: the browser will choose the most suitable one. It can also be the destination for streamed media, using a MediaStream."><audio> elements.

 

method createMediaElementSource(parmyMediaElement: dynamic): MediaElementAudioSourceNode

 

MediaElementAudioSourceNode createMediaElementSource(dynamic parmyMediaElement)

 

func createMediaElementSource(_ parmyMediaElement: dynamic) -> MediaElementAudioSourceNode

 

MediaElementAudioSourceNode createMediaElementSource(dynamic parmyMediaElement)

 

Function createMediaElementSource(parmyMediaElement As dynamic) As MediaElementAudioSourceNode

Parameters:

  • parmyMediaElement:

createMediaStreamDestination

Creates a MediaStreamAudioDestinationNode associated with a MediaStream representing an audio stream which may be stored in a local file or sent to another computer.

 

method createMediaStreamDestination: MediaStream

 

MediaStream createMediaStreamDestination()

 

func createMediaStreamDestination() -> MediaStream

 

MediaStream createMediaStreamDestination()

 

Function createMediaStreamDestination() As MediaStream

createMediaStreamSource

Creates a MediaStreamAudioSourceNode associated with a MediaStream representing an audio stream which may come from the local computer microphone or other sources.

 

method createMediaStreamSource(parstream: dynamic): MediaStream

 

MediaStream createMediaStreamSource(dynamic parstream)

 

func createMediaStreamSource(_ parstream: dynamic) -> MediaStream

 

MediaStream createMediaStreamSource(dynamic parstream)

 

Function createMediaStreamSource(parstream As dynamic) As MediaStream

Parameters:

  • parstream:

createMediaStreamTrackSource

Creates a MediaStreamTrackAudioSourceNode associated with a MediaStream representing an media stream track.

 

method createMediaStreamTrackSource(partrack: dynamic): MediaStream

 

MediaStream createMediaStreamTrackSource(dynamic partrack)

 

func createMediaStreamTrackSource(_ partrack: dynamic) -> MediaStream

 

MediaStream createMediaStreamTrackSource(dynamic partrack)

 

Function createMediaStreamTrackSource(partrack As dynamic) As MediaStream

Parameters:

  • partrack:

createOscillator    (declared in BaseAudioContext)

Creates an OscillatorNode, a source representing a periodic waveform. It basically generates a tone.

 

method createOscillator: OscillatorNode

 

OscillatorNode createOscillator()

 

func createOscillator() -> OscillatorNode

 

OscillatorNode createOscillator()

 

Function createOscillator() As OscillatorNode

createPanner    (declared in BaseAudioContext)

Creates a PannerNode, which is used to spatialise an incoming audio stream in 3D space.

 

method createPanner: PannerNode

 

PannerNode createPanner()

 

func createPanner() -> PannerNode

 

PannerNode createPanner()

 

Function createPanner() As PannerNode

createPeriodicWave    (declared in BaseAudioContext)

Creates a PeriodicWave, used to define a periodic waveform that can be used to determine the output of an OscillatorNode.

 

method createPeriodicWave(parreal: dynamic; parimag: dynamic): PeriodicWave

 

PeriodicWave createPeriodicWave(dynamic parreal, dynamic parimag)

 

func createPeriodicWave(_ parreal: dynamic, _ parimag: dynamic) -> PeriodicWave

 

PeriodicWave createPeriodicWave(dynamic parreal, dynamic parimag)

 

Function createPeriodicWave(parreal As dynamic, parimag As dynamic) As PeriodicWave

Parameters:

  • parreal:
  • parimag:

createStereoPanner    (declared in BaseAudioContext)

Creates a StereoPannerNode, which can be used to apply stereo panning to an audio source.

 

method createStereoPanner: StereoPannerNode

 

StereoPannerNode createStereoPanner()

 

func createStereoPanner() -> StereoPannerNode

 

StereoPannerNode createStereoPanner()

 

Function createStereoPanner() As StereoPannerNode

createWaveShaper    (declared in BaseAudioContext)

Creates a WaveShaperNode, which is used to implement non-linear distortion effects.

 

method createWaveShaper: WaveShaperNode

 

WaveShaperNode createWaveShaper()

 

func createWaveShaper() -> WaveShaperNode

 

WaveShaperNode createWaveShaper()

 

Function createWaveShaper() As WaveShaperNode

currentTime    (declared in BaseAudioContext)

 

property currentTime: DateTime read;

 

DateTime currentTime { get; }

 

var currentTime: DateTime { get{} }

 

DateTime currentTime { __get; }

 

ReadOnly Property currentTime() As DateTime

decodeAudioData    (declared in BaseAudioContext)

Asynchronously decodes audio file data contained in an ArrayBuffer. In this case, the ArrayBuffer is usually loaded from an XMLHttpRequest's response attribute after setting the responseType to arraybuffer. This method only works on complete files, not fragments of audio files.

 

method decodeAudioData(parArrayBuffer: dynamic; parDecodeSuccessCallback: dynamic; parDecodeErrorCallback: dynamic): array of Byte

 

Byte[] decodeAudioData(dynamic parArrayBuffer, dynamic parDecodeSuccessCallback, dynamic parDecodeErrorCallback)

 

func decodeAudioData(_ parArrayBuffer: dynamic, _ parDecodeSuccessCallback: dynamic, _ parDecodeErrorCallback: dynamic) -> Byte...

 

Byte[] decodeAudioData(dynamic parArrayBuffer, dynamic parDecodeSuccessCallback, dynamic parDecodeErrorCallback)

 

Function decodeAudioData(parArrayBuffer As dynamic, parDecodeSuccessCallback As dynamic, parDecodeErrorCallback As dynamic) As Byte()

Parameters:

  • parArrayBuffer:
  • parDecodeSuccessCallback:
  • parDecodeErrorCallback:

destination    (declared in BaseAudioContext)

 

property destination: AudioDestinationNode read;

 

AudioDestinationNode destination { get; }

 

var destination: AudioDestinationNode { get{} }

 

AudioDestinationNode destination { __get; }

 

ReadOnly Property destination() As AudioDestinationNode

getOutputTimestamp

Returns a new AudioTimestamp object containing two audio timestamp values relating to the current audio context.

 

method getOutputTimestamp: DateTime

 

DateTime getOutputTimestamp()

 

func getOutputTimestamp() -> DateTime

 

DateTime getOutputTimestamp()

 

Function getOutputTimestamp() As DateTime

listener    (declared in BaseAudioContext)

 

property listener: AudioListener read;

 

AudioListener listener { get; }

 

var listener: AudioListener { get{} }

 

AudioListener listener { __get; }

 

ReadOnly Property listener() As AudioListener

onstatechange    (declared in BaseAudioContext)

 

property onstatechange: EventListener read write;

 

EventListener onstatechange { get; set; }

 

var onstatechange: EventListener { get{} set{} }

 

EventListener onstatechange { __get; __set; }

 

Property onstatechange() As EventListener

sampleRate    (declared in BaseAudioContext)

 

property sampleRate: Double read;

 

Double sampleRate { get; }

 

var sampleRate: Double { get{} }

 

Double sampleRate { __get; }

 

ReadOnly Property sampleRate() As Double

state    (declared in BaseAudioContext)

 

property state: String read;

 

String state { get; }

 

var state: String { get{} }

 

String state { __get; }

 

ReadOnly Property state() As String

 

currentTime    (declared in BaseAudioContext)

 

property currentTime: DateTime read;

 

DateTime currentTime { get; }

 

var currentTime: DateTime { get{} }

 

DateTime currentTime { __get; }

 

ReadOnly Property currentTime() As DateTime

destination    (declared in BaseAudioContext)

 

property destination: AudioDestinationNode read;

 

AudioDestinationNode destination { get; }

 

var destination: AudioDestinationNode { get{} }

 

AudioDestinationNode destination { __get; }

 

ReadOnly Property destination() As AudioDestinationNode

listener    (declared in BaseAudioContext)

 

property listener: AudioListener read;

 

AudioListener listener { get; }

 

var listener: AudioListener { get{} }

 

AudioListener listener { __get; }

 

ReadOnly Property listener() As AudioListener

onstatechange    (declared in BaseAudioContext)

 

property onstatechange: EventListener read write;

 

EventListener onstatechange { get; set; }

 

var onstatechange: EventListener { get{} set{} }

 

EventListener onstatechange { __get; __set; }

 

Property onstatechange() As EventListener

sampleRate    (declared in BaseAudioContext)

 

property sampleRate: Double read;

 

Double sampleRate { get; }

 

var sampleRate: Double { get{} }

 

Double sampleRate { __get; }

 

ReadOnly Property sampleRate() As Double

state    (declared in BaseAudioContext)

 

property state: String read;

 

String state { get; }

 

var state: String { get{} }

 

String state { __get; }

 

ReadOnly Property state() As String

 

createAnalyser    (declared in BaseAudioContext)

Creates an AnalyserNode, which can be used to expose audio time and frequency data and for example to create data visualisations.

 

method createAnalyser: AnalyserNode

 

AnalyserNode createAnalyser()

 

func createAnalyser() -> AnalyserNode

 

AnalyserNode createAnalyser()

 

Function createAnalyser() As AnalyserNode

createBiquadFilter    (declared in BaseAudioContext)

Creates a BiquadFilterNode, which represents a second order filter configurable as several different common filter types: high-pass, low-pass, band-pass, etc

 

method createBiquadFilter: BiquadFilterNode

 

BiquadFilterNode createBiquadFilter()

 

func createBiquadFilter() -> BiquadFilterNode

 

BiquadFilterNode createBiquadFilter()

 

Function createBiquadFilter() As BiquadFilterNode

createBufferSource    (declared in BaseAudioContext)

Creates an AudioBufferSourceNode, which can be used to play and manipulate audio data contained within an AudioBuffer object. AudioBuffers are created using AudioContext.createBuffer or returned by AudioContext.decodeAudioData when it successfully decodes an audio track.

 

method createBufferSource: AudioBufferSourceNode

 

AudioBufferSourceNode createBufferSource()

 

func createBufferSource() -> AudioBufferSourceNode

 

AudioBufferSourceNode createBufferSource()

 

Function createBufferSource() As AudioBufferSourceNode

createChannelMerger    (declared in BaseAudioContext)

Creates a ChannelMergerNode, which is used to combine channels from multiple audio streams into a single audio stream.

 

method createChannelMerger(parnumberOfInputs: dynamic): ChannelMergerNode

 

ChannelMergerNode createChannelMerger(dynamic parnumberOfInputs)

 

func createChannelMerger(_ parnumberOfInputs: dynamic) -> ChannelMergerNode

 

ChannelMergerNode createChannelMerger(dynamic parnumberOfInputs)

 

Function createChannelMerger(parnumberOfInputs As dynamic) As ChannelMergerNode

Parameters:

  • parnumberOfInputs:

createChannelSplitter    (declared in BaseAudioContext)

Creates a ChannelSplitterNode, which is used to access the individual channels of an audio stream and process them separately.

 

method createChannelSplitter(parnumberOfOutputs: dynamic): ChannelSplitterNode

 

ChannelSplitterNode createChannelSplitter(dynamic parnumberOfOutputs)

 

func createChannelSplitter(_ parnumberOfOutputs: dynamic) -> ChannelSplitterNode

 

ChannelSplitterNode createChannelSplitter(dynamic parnumberOfOutputs)

 

Function createChannelSplitter(parnumberOfOutputs As dynamic) As ChannelSplitterNode

Parameters:

  • parnumberOfOutputs:

createConstantSource    (declared in BaseAudioContext)

Creates a ConstantSourceNode object, which is an audio source that continuously outputs a monaural (one-channel) sound signal whose samples all have the same value.

 

method createConstantSource: ConstantSourceNode

 

ConstantSourceNode createConstantSource()

 

func createConstantSource() -> ConstantSourceNode

 

ConstantSourceNode createConstantSource()

 

Function createConstantSource() As ConstantSourceNode

createConvolver    (declared in BaseAudioContext)

Creates a ConvolverNode, which can be used to apply convolution effects to your audio graph, for example a reverberation effect.

 

method createConvolver: ConvolverNode

 

ConvolverNode createConvolver()

 

func createConvolver() -> ConvolverNode

 

ConvolverNode createConvolver()

 

Function createConvolver() As ConvolverNode

createDelay    (declared in BaseAudioContext)

Creates a DelayNode, which is used to delay the incoming audio signal by a certain amount. This node is also useful to create feedback loops in a Web Audio API graph.

 

method createDelay(parmaxDelayTime: dynamic): DelayNode

 

DelayNode createDelay(dynamic parmaxDelayTime)

 

func createDelay(_ parmaxDelayTime: dynamic) -> DelayNode

 

DelayNode createDelay(dynamic parmaxDelayTime)

 

Function createDelay(parmaxDelayTime As dynamic) As DelayNode

Parameters:

  • parmaxDelayTime:

createDynamicsCompressor    (declared in BaseAudioContext)

Creates a DynamicsCompressorNode, which can be used to apply acoustic compression to an audio signal.

 

method createDynamicsCompressor: DynamicsCompressorNode

 

DynamicsCompressorNode createDynamicsCompressor()

 

func createDynamicsCompressor() -> DynamicsCompressorNode

 

DynamicsCompressorNode createDynamicsCompressor()

 

Function createDynamicsCompressor() As DynamicsCompressorNode

createGain    (declared in BaseAudioContext)

Creates a GainNode, which can be used to control the overall volume of the audio graph.

 

method createGain: GainNode

 

GainNode createGain()

 

func createGain() -> GainNode

 

GainNode createGain()

 

Function createGain() As GainNode

createIIRFilter    (declared in BaseAudioContext)

Creates an IIRFilterNode, which represents a second order filter configurable as several different common filter types.

 

method createIIRFilter(parfeedforward: dynamic; parfeedback: dynamic): IIRFilterNode

 

IIRFilterNode createIIRFilter(dynamic parfeedforward, dynamic parfeedback)

 

func createIIRFilter(_ parfeedforward: dynamic, _ parfeedback: dynamic) -> IIRFilterNode

 

IIRFilterNode createIIRFilter(dynamic parfeedforward, dynamic parfeedback)

 

Function createIIRFilter(parfeedforward As dynamic, parfeedback As dynamic) As IIRFilterNode

Parameters:

  • parfeedforward:
  • parfeedback:

createMediaElementSource

Creates a MediaElementAudioSourceNode associated with an HTMLMediaElement. This can be used to play and manipulate audio from ) embeds a media player which supports video playback into the document. You can use <video> for audio content as well, but the <audio> element may provide a more appropriate user experience."><video> or element is used to embed sound content in documents. It may contain one or more audio sources, represented using the src attribute or the <source> element: the browser will choose the most suitable one. It can also be the destination for streamed media, using a MediaStream."><audio> elements.

 

method createMediaElementSource(parmyMediaElement: dynamic): MediaElementAudioSourceNode

 

MediaElementAudioSourceNode createMediaElementSource(dynamic parmyMediaElement)

 

func createMediaElementSource(_ parmyMediaElement: dynamic) -> MediaElementAudioSourceNode

 

MediaElementAudioSourceNode createMediaElementSource(dynamic parmyMediaElement)

 

Function createMediaElementSource(parmyMediaElement As dynamic) As MediaElementAudioSourceNode

Parameters:

  • parmyMediaElement:

createMediaStreamDestination

Creates a MediaStreamAudioDestinationNode associated with a MediaStream representing an audio stream which may be stored in a local file or sent to another computer.

 

method createMediaStreamDestination: MediaStream

 

MediaStream createMediaStreamDestination()

 

func createMediaStreamDestination() -> MediaStream

 

MediaStream createMediaStreamDestination()

 

Function createMediaStreamDestination() As MediaStream

createMediaStreamSource

Creates a MediaStreamAudioSourceNode associated with a MediaStream representing an audio stream which may come from the local computer microphone or other sources.

 

method createMediaStreamSource(parstream: dynamic): MediaStream

 

MediaStream createMediaStreamSource(dynamic parstream)

 

func createMediaStreamSource(_ parstream: dynamic) -> MediaStream

 

MediaStream createMediaStreamSource(dynamic parstream)

 

Function createMediaStreamSource(parstream As dynamic) As MediaStream

Parameters:

  • parstream:

createMediaStreamTrackSource

Creates a MediaStreamTrackAudioSourceNode associated with a MediaStream representing an media stream track.

 

method createMediaStreamTrackSource(partrack: dynamic): MediaStream

 

MediaStream createMediaStreamTrackSource(dynamic partrack)

 

func createMediaStreamTrackSource(_ partrack: dynamic) -> MediaStream

 

MediaStream createMediaStreamTrackSource(dynamic partrack)

 

Function createMediaStreamTrackSource(partrack As dynamic) As MediaStream

Parameters:

  • partrack:

createOscillator    (declared in BaseAudioContext)

Creates an OscillatorNode, a source representing a periodic waveform. It basically generates a tone.

 

method createOscillator: OscillatorNode

 

OscillatorNode createOscillator()

 

func createOscillator() -> OscillatorNode

 

OscillatorNode createOscillator()

 

Function createOscillator() As OscillatorNode

createPanner    (declared in BaseAudioContext)

Creates a PannerNode, which is used to spatialise an incoming audio stream in 3D space.

 

method createPanner: PannerNode

 

PannerNode createPanner()

 

func createPanner() -> PannerNode

 

PannerNode createPanner()

 

Function createPanner() As PannerNode

createPeriodicWave    (declared in BaseAudioContext)

Creates a PeriodicWave, used to define a periodic waveform that can be used to determine the output of an OscillatorNode.

 

method createPeriodicWave(parreal: dynamic; parimag: dynamic): PeriodicWave

 

PeriodicWave createPeriodicWave(dynamic parreal, dynamic parimag)

 

func createPeriodicWave(_ parreal: dynamic, _ parimag: dynamic) -> PeriodicWave

 

PeriodicWave createPeriodicWave(dynamic parreal, dynamic parimag)

 

Function createPeriodicWave(parreal As dynamic, parimag As dynamic) As PeriodicWave

Parameters:

  • parreal:
  • parimag:

createStereoPanner    (declared in BaseAudioContext)

Creates a StereoPannerNode, which can be used to apply stereo panning to an audio source.

 

method createStereoPanner: StereoPannerNode

 

StereoPannerNode createStereoPanner()

 

func createStereoPanner() -> StereoPannerNode

 

StereoPannerNode createStereoPanner()

 

Function createStereoPanner() As StereoPannerNode

createWaveShaper    (declared in BaseAudioContext)

Creates a WaveShaperNode, which is used to implement non-linear distortion effects.

 

method createWaveShaper: WaveShaperNode

 

WaveShaperNode createWaveShaper()

 

func createWaveShaper() -> WaveShaperNode

 

WaveShaperNode createWaveShaper()

 

Function createWaveShaper() As WaveShaperNode

decodeAudioData    (declared in BaseAudioContext)

Asynchronously decodes audio file data contained in an ArrayBuffer. In this case, the ArrayBuffer is usually loaded from an XMLHttpRequest's response attribute after setting the responseType to arraybuffer. This method only works on complete files, not fragments of audio files.

 

method decodeAudioData(parArrayBuffer: dynamic; parDecodeSuccessCallback: dynamic; parDecodeErrorCallback: dynamic): array of Byte

 

Byte[] decodeAudioData(dynamic parArrayBuffer, dynamic parDecodeSuccessCallback, dynamic parDecodeErrorCallback)

 

func decodeAudioData(_ parArrayBuffer: dynamic, _ parDecodeSuccessCallback: dynamic, _ parDecodeErrorCallback: dynamic) -> Byte...

 

Byte[] decodeAudioData(dynamic parArrayBuffer, dynamic parDecodeSuccessCallback, dynamic parDecodeErrorCallback)

 

Function decodeAudioData(parArrayBuffer As dynamic, parDecodeSuccessCallback As dynamic, parDecodeErrorCallback As dynamic) As Byte()

Parameters:

  • parArrayBuffer:
  • parDecodeSuccessCallback:
  • parDecodeErrorCallback:

getOutputTimestamp

Returns a new AudioTimestamp object containing two audio timestamp values relating to the current audio context.

 

method getOutputTimestamp: DateTime

 

DateTime getOutputTimestamp()

 

func getOutputTimestamp() -> DateTime

 

DateTime getOutputTimestamp()

 

Function getOutputTimestamp() As DateTime