| Index: Source/modules/webaudio/AudioContext.cpp
|
| diff --git a/Source/modules/webaudio/AudioContext.cpp b/Source/modules/webaudio/AudioContext.cpp
|
| index 2997ae52904e6ee5be5c55cfdd019524bfd78cce..508d8c941c023d35e407c291f9e1446681b15589 100644
|
| --- a/Source/modules/webaudio/AudioContext.cpp
|
| +++ b/Source/modules/webaudio/AudioContext.cpp
|
| @@ -89,11 +89,11 @@ bool AudioContext::isSampleRateRangeGood(float sampleRate)
|
| const unsigned MaxHardwareContexts = 4;
|
| unsigned AudioContext::s_hardwareContextCount = 0;
|
|
|
| -PassRefPtr<AudioContext> AudioContext::create(Document& document, ExceptionState& es)
|
| +PassRefPtr<AudioContext> AudioContext::create(Document& document, ExceptionState& exceptionState)
|
| {
|
| ASSERT(isMainThread());
|
| if (s_hardwareContextCount >= MaxHardwareContexts) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| SyntaxError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioContext",
|
| @@ -106,10 +106,10 @@ PassRefPtr<AudioContext> AudioContext::create(Document& document, ExceptionState
|
| return audioContext.release();
|
| }
|
|
|
| -PassRefPtr<AudioContext> AudioContext::create(Document& document, unsigned numberOfChannels, size_t numberOfFrames, float sampleRate, ExceptionState& es)
|
| +PassRefPtr<AudioContext> AudioContext::create(Document& document, unsigned numberOfChannels, size_t numberOfFrames, float sampleRate, ExceptionState& exceptionState)
|
| {
|
| document.addConsoleMessage(JSMessageSource, WarningMessageLevel, "Deprecated AudioContext constructor: use OfflineAudioContext instead");
|
| - return OfflineAudioContext::create(&document, numberOfChannels, numberOfFrames, sampleRate, es);
|
| + return OfflineAudioContext::create(&document, numberOfChannels, numberOfFrames, sampleRate, exceptionState);
|
| }
|
|
|
| // Constructor for rendering to the audio hardware.
|
| @@ -296,18 +296,18 @@ void AudioContext::stop()
|
| callOnMainThread(stopDispatch, this);
|
| }
|
|
|
| -PassRefPtr<AudioBuffer> AudioContext::createBuffer(unsigned numberOfChannels, size_t numberOfFrames, float sampleRate, ExceptionState& es)
|
| +PassRefPtr<AudioBuffer> AudioContext::createBuffer(unsigned numberOfChannels, size_t numberOfFrames, float sampleRate, ExceptionState& exceptionState)
|
| {
|
| RefPtr<AudioBuffer> audioBuffer = AudioBuffer::create(numberOfChannels, numberOfFrames, sampleRate);
|
| if (!audioBuffer.get()) {
|
| if (numberOfChannels > AudioContext::maxNumberOfChannels()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| NotSupportedError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioBuffer",
|
| "requested number of channels (" + String::number(numberOfChannels) + ") exceeds maximum (" + String::number(AudioContext::maxNumberOfChannels()) + ")"));
|
| } else if (sampleRate < AudioBuffer::minAllowedSampleRate() || sampleRate > AudioBuffer::maxAllowedSampleRate()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| NotSupportedError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioBuffer",
|
| @@ -316,13 +316,13 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(unsigned numberOfChannels, si
|
| + String::number(AudioBuffer::minAllowedSampleRate())
|
| + "-" + String::number(AudioBuffer::maxAllowedSampleRate()) + " Hz"));
|
| } else if (!numberOfFrames) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| NotSupportedError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioBuffer",
|
| "number of frames must be greater than 0."));
|
| } else {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| NotSupportedError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioBuffer",
|
| @@ -336,11 +336,11 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(unsigned numberOfChannels, si
|
| return audioBuffer;
|
| }
|
|
|
| -PassRefPtr<AudioBuffer> AudioContext::createBuffer(ArrayBuffer* arrayBuffer, bool mixToMono, ExceptionState& es)
|
| +PassRefPtr<AudioBuffer> AudioContext::createBuffer(ArrayBuffer* arrayBuffer, bool mixToMono, ExceptionState& exceptionState)
|
| {
|
| ASSERT(arrayBuffer);
|
| if (!arrayBuffer) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| SyntaxError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioBuffer",
|
| @@ -350,7 +350,7 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(ArrayBuffer* arrayBuffer, boo
|
|
|
| RefPtr<AudioBuffer> audioBuffer = AudioBuffer::createFromAudioFileData(arrayBuffer->data(), arrayBuffer->byteLength(), mixToMono, sampleRate());
|
| if (!audioBuffer.get()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| SyntaxError,
|
| ExceptionMessages::failedToConstruct(
|
| "AudioBuffer",
|
| @@ -361,10 +361,10 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(ArrayBuffer* arrayBuffer, boo
|
| return audioBuffer;
|
| }
|
|
|
| -void AudioContext::decodeAudioData(ArrayBuffer* audioData, PassRefPtr<AudioBufferCallback> successCallback, PassRefPtr<AudioBufferCallback> errorCallback, ExceptionState& es)
|
| +void AudioContext::decodeAudioData(ArrayBuffer* audioData, PassRefPtr<AudioBufferCallback> successCallback, PassRefPtr<AudioBufferCallback> errorCallback, ExceptionState& exceptionState)
|
| {
|
| if (!audioData) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| SyntaxError,
|
| ExceptionMessages::failedToExecute(
|
| "decodeAudioData",
|
| @@ -388,10 +388,10 @@ PassRefPtr<AudioBufferSourceNode> AudioContext::createBufferSource()
|
| return node;
|
| }
|
|
|
| -PassRefPtr<MediaElementAudioSourceNode> AudioContext::createMediaElementSource(HTMLMediaElement* mediaElement, ExceptionState& es)
|
| +PassRefPtr<MediaElementAudioSourceNode> AudioContext::createMediaElementSource(HTMLMediaElement* mediaElement, ExceptionState& exceptionState)
|
| {
|
| if (!mediaElement) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| InvalidStateError,
|
| ExceptionMessages::failedToConstruct(
|
| "MediaElementAudioSourceNode",
|
| @@ -404,7 +404,7 @@ PassRefPtr<MediaElementAudioSourceNode> AudioContext::createMediaElementSource(H
|
|
|
| // First check if this media element already has a source node.
|
| if (mediaElement->audioSourceNode()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| InvalidStateError,
|
| ExceptionMessages::failedToConstruct(
|
| "MediaElementAudioSourceNode",
|
| @@ -420,10 +420,10 @@ PassRefPtr<MediaElementAudioSourceNode> AudioContext::createMediaElementSource(H
|
| return node;
|
| }
|
|
|
| -PassRefPtr<MediaStreamAudioSourceNode> AudioContext::createMediaStreamSource(MediaStream* mediaStream, ExceptionState& es)
|
| +PassRefPtr<MediaStreamAudioSourceNode> AudioContext::createMediaStreamSource(MediaStream* mediaStream, ExceptionState& exceptionState)
|
| {
|
| if (!mediaStream) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| InvalidStateError,
|
| ExceptionMessages::failedToConstruct(
|
| "MediaStreamAudioSourceNode",
|
| @@ -463,25 +463,25 @@ PassRefPtr<MediaStreamAudioDestinationNode> AudioContext::createMediaStreamDesti
|
| return MediaStreamAudioDestinationNode::create(this, 1);
|
| }
|
|
|
| -PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(ExceptionState& es)
|
| +PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(ExceptionState& exceptionState)
|
| {
|
| // Set number of input/output channels to stereo by default.
|
| - return createScriptProcessor(0, 2, 2, es);
|
| + return createScriptProcessor(0, 2, 2, exceptionState);
|
| }
|
|
|
| -PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t bufferSize, ExceptionState& es)
|
| +PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t bufferSize, ExceptionState& exceptionState)
|
| {
|
| // Set number of input/output channels to stereo by default.
|
| - return createScriptProcessor(bufferSize, 2, 2, es);
|
| + return createScriptProcessor(bufferSize, 2, 2, exceptionState);
|
| }
|
|
|
| -PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t bufferSize, size_t numberOfInputChannels, ExceptionState& es)
|
| +PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t bufferSize, size_t numberOfInputChannels, ExceptionState& exceptionState)
|
| {
|
| // Set number of output channels to stereo by default.
|
| - return createScriptProcessor(bufferSize, numberOfInputChannels, 2, es);
|
| + return createScriptProcessor(bufferSize, numberOfInputChannels, 2, exceptionState);
|
| }
|
|
|
| -PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t bufferSize, size_t numberOfInputChannels, size_t numberOfOutputChannels, ExceptionState& es)
|
| +PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t bufferSize, size_t numberOfInputChannels, size_t numberOfOutputChannels, ExceptionState& exceptionState)
|
| {
|
| ASSERT(isMainThread());
|
| lazyInitialize();
|
| @@ -489,13 +489,13 @@ PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t buffe
|
|
|
| if (!node.get()) {
|
| if (!numberOfInputChannels && !numberOfOutputChannels) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "ScriptProcessorNode",
|
| "number of input channels and output channels cannot both be zero."));
|
| } else if (numberOfInputChannels > AudioContext::maxNumberOfChannels()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "ScriptProcessorNode",
|
| @@ -503,7 +503,7 @@ PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t buffe
|
| + ") exceeds maximum ("
|
| + String::number(AudioContext::maxNumberOfChannels()) + ")."));
|
| } else if (numberOfOutputChannels > AudioContext::maxNumberOfChannels()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "ScriptProcessorNode",
|
| @@ -511,7 +511,7 @@ PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t buffe
|
| + ") exceeds maximum ("
|
| + String::number(AudioContext::maxNumberOfChannels()) + ")."));
|
| } else {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "ScriptProcessorNode",
|
| @@ -574,29 +574,29 @@ PassRefPtr<GainNode> AudioContext::createGain()
|
| return GainNode::create(this, m_destinationNode->sampleRate());
|
| }
|
|
|
| -PassRefPtr<DelayNode> AudioContext::createDelay(ExceptionState& es)
|
| +PassRefPtr<DelayNode> AudioContext::createDelay(ExceptionState& exceptionState)
|
| {
|
| const double defaultMaxDelayTime = 1;
|
| - return createDelay(defaultMaxDelayTime, es);
|
| + return createDelay(defaultMaxDelayTime, exceptionState);
|
| }
|
|
|
| -PassRefPtr<DelayNode> AudioContext::createDelay(double maxDelayTime, ExceptionState& es)
|
| +PassRefPtr<DelayNode> AudioContext::createDelay(double maxDelayTime, ExceptionState& exceptionState)
|
| {
|
| ASSERT(isMainThread());
|
| lazyInitialize();
|
| - RefPtr<DelayNode> node = DelayNode::create(this, m_destinationNode->sampleRate(), maxDelayTime, es);
|
| - if (es.hadException())
|
| + RefPtr<DelayNode> node = DelayNode::create(this, m_destinationNode->sampleRate(), maxDelayTime, exceptionState);
|
| + if (exceptionState.hadException())
|
| return 0;
|
| return node;
|
| }
|
|
|
| -PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(ExceptionState& es)
|
| +PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(ExceptionState& exceptionState)
|
| {
|
| const unsigned ChannelSplitterDefaultNumberOfOutputs = 6;
|
| - return createChannelSplitter(ChannelSplitterDefaultNumberOfOutputs, es);
|
| + return createChannelSplitter(ChannelSplitterDefaultNumberOfOutputs, exceptionState);
|
| }
|
|
|
| -PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(size_t numberOfOutputs, ExceptionState& es)
|
| +PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(size_t numberOfOutputs, ExceptionState& exceptionState)
|
| {
|
| ASSERT(isMainThread());
|
| lazyInitialize();
|
| @@ -604,7 +604,7 @@ PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(size_t numbe
|
| RefPtr<ChannelSplitterNode> node = ChannelSplitterNode::create(this, m_destinationNode->sampleRate(), numberOfOutputs);
|
|
|
| if (!node.get()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "ChannelSplitterNode",
|
| @@ -617,13 +617,13 @@ PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(size_t numbe
|
| return node;
|
| }
|
|
|
| -PassRefPtr<ChannelMergerNode> AudioContext::createChannelMerger(ExceptionState& es)
|
| +PassRefPtr<ChannelMergerNode> AudioContext::createChannelMerger(ExceptionState& exceptionState)
|
| {
|
| const unsigned ChannelMergerDefaultNumberOfInputs = 6;
|
| - return createChannelMerger(ChannelMergerDefaultNumberOfInputs, es);
|
| + return createChannelMerger(ChannelMergerDefaultNumberOfInputs, exceptionState);
|
| }
|
|
|
| -PassRefPtr<ChannelMergerNode> AudioContext::createChannelMerger(size_t numberOfInputs, ExceptionState& es)
|
| +PassRefPtr<ChannelMergerNode> AudioContext::createChannelMerger(size_t numberOfInputs, ExceptionState& exceptionState)
|
| {
|
| ASSERT(isMainThread());
|
| lazyInitialize();
|
| @@ -631,7 +631,7 @@ PassRefPtr<ChannelMergerNode> AudioContext::createChannelMerger(size_t numberOfI
|
| RefPtr<ChannelMergerNode> node = ChannelMergerNode::create(this, m_destinationNode->sampleRate(), numberOfInputs);
|
|
|
| if (!node.get()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "ChannelMergerNode",
|
| @@ -658,12 +658,12 @@ PassRefPtr<OscillatorNode> AudioContext::createOscillator()
|
| return node;
|
| }
|
|
|
| -PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Float32Array* imag, ExceptionState& es)
|
| +PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Float32Array* imag, ExceptionState& exceptionState)
|
| {
|
| ASSERT(isMainThread());
|
|
|
| if (!real) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| SyntaxError,
|
| ExceptionMessages::failedToConstruct(
|
| "PeriodicWave",
|
| @@ -672,7 +672,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| }
|
|
|
| if (!imag) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| SyntaxError,
|
| ExceptionMessages::failedToConstruct(
|
| "PeriodicWave",
|
| @@ -681,7 +681,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| }
|
|
|
| if (real->length() != imag->length()) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "PeriodicWave",
|
| @@ -692,7 +692,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| }
|
|
|
| if (real->length() > 4096) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "PeriodicWave",
|
| @@ -702,7 +702,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| }
|
|
|
| if (imag->length() > 4096) {
|
| - es.throwDOMException(
|
| + exceptionState.throwDOMException(
|
| IndexSizeError,
|
| ExceptionMessages::failedToConstruct(
|
| "PeriodicWave",
|
|
|