AudioContext interface represents an audio-processing graph built from audio modules linked together, each represented by an
An audio context controls both the creation of the nodes it contains and the execution of the audio processing, or decoding. You need to create an
AudioContext before you do anything else, as everything happens inside a context. It's recommended to create one AudioContext and reuse it instead of initializing a new one each time, and it's OK to use a single
AudioContext for several different audio sources and pipeline concurrently.
Creates and returns a new
Also inherits properties from its parent interface,
AudioContext.baseLatencyRead only Experimental
AudioContext.outputLatencyRead only Experimental
Returns an estimation of the output latency of the current audio context.
Also inherits methods from its parent interface,
Closes the audio context, releasing any system audio resources that it uses.
Returns a new
AudioTimestampobject containing two audio timestamp values relating to the current audio context.
Resumes the progression of time in an audio context that has previously been suspended/paused.
Suspends the progression of time in the audio context, temporarily halting audio hardware access and reducing CPU/battery usage in the process.
Basic audio context declaration:
var audioCtx = new AudioContext();
Cross browser variant:
var AudioContext = window.AudioContext || window.webkitAudioContext; var audioCtx = new AudioContext(); var oscillatorNode = audioCtx.createOscillator(); var gainNode = audioCtx.createGain(); var finish = audioCtx.destination; // etc.
|Web Audio API |
BCD tables only load in the browser