| Index: Source/modules/webaudio/AudioContext.cpp
|
| diff --git a/Source/modules/webaudio/AudioContext.cpp b/Source/modules/webaudio/AudioContext.cpp
|
| index b0d76567e71d15e96ec2c170e9c24d724ce706c2..c244bed068a7de453e29134918ba8878e6637675 100644
|
| --- a/Source/modules/webaudio/AudioContext.cpp
|
| +++ b/Source/modules/webaudio/AudioContext.cpp
|
| @@ -96,7 +96,7 @@ PassRefPtr<AudioContext> AudioContext::create(Document& document, ExceptionState
|
| exceptionState.throwDOMException(
|
| SyntaxError,
|
| "number of hardware contexts reached maximum (" + String::number(MaxHardwareContexts) + ").");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| RefPtr<AudioContext> audioContext(adoptRef(new AudioContext(&document)));
|
| @@ -116,7 +116,7 @@ AudioContext::AudioContext(Document* document)
|
| , m_isStopScheduled(false)
|
| , m_isInitialized(false)
|
| , m_isAudioThreadFinished(false)
|
| - , m_destinationNode(0)
|
| + , m_destinationNode(nullptr)
|
| , m_isDeletionScheduled(false)
|
| , m_automaticPullNodesNeedUpdating(false)
|
| , m_connectionCount(0)
|
| @@ -142,7 +142,7 @@ AudioContext::AudioContext(Document* document, unsigned numberOfChannels, size_t
|
| , m_isStopScheduled(false)
|
| , m_isInitialized(false)
|
| , m_isAudioThreadFinished(false)
|
| - , m_destinationNode(0)
|
| + , m_destinationNode(nullptr)
|
| , m_automaticPullNodesNeedUpdating(false)
|
| , m_connectionCount(0)
|
| , m_audioThread(0)
|
| @@ -320,7 +320,7 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(unsigned numberOfChannels, si
|
| + " channel(s) of " + String::number(numberOfFrames)
|
| + " frames each.");
|
| }
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| return audioBuffer;
|
| @@ -333,7 +333,7 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(ArrayBuffer* arrayBuffer, boo
|
| exceptionState.throwDOMException(
|
| SyntaxError,
|
| "invalid ArrayBuffer.");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| RefPtr<AudioBuffer> audioBuffer = AudioBuffer::createFromAudioFileData(arrayBuffer->data(), arrayBuffer->byteLength(), mixToMono, sampleRate());
|
| @@ -341,7 +341,7 @@ PassRefPtr<AudioBuffer> AudioContext::createBuffer(ArrayBuffer* arrayBuffer, boo
|
| exceptionState.throwDOMException(
|
| SyntaxError,
|
| "invalid audio data in ArrayBuffer.");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| return audioBuffer;
|
| @@ -377,7 +377,7 @@ PassRefPtr<MediaElementAudioSourceNode> AudioContext::createMediaElementSource(H
|
| exceptionState.throwDOMException(
|
| InvalidStateError,
|
| "invalid HTMLMedialElement.");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| ASSERT(isMainThread());
|
| @@ -388,7 +388,7 @@ PassRefPtr<MediaElementAudioSourceNode> AudioContext::createMediaElementSource(H
|
| exceptionState.throwDOMException(
|
| InvalidStateError,
|
| "invalid HTMLMediaElement.");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| RefPtr<MediaElementAudioSourceNode> node = MediaElementAudioSourceNode::create(this, mediaElement);
|
| @@ -405,7 +405,7 @@ PassRefPtr<MediaStreamAudioSourceNode> AudioContext::createMediaStreamSource(Med
|
| exceptionState.throwDOMException(
|
| InvalidStateError,
|
| "invalid MediaStream source");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| ASSERT(isMainThread());
|
| @@ -488,7 +488,7 @@ PassRefPtr<ScriptProcessorNode> AudioContext::createScriptProcessor(size_t buffe
|
| "buffer size (" + String::number(bufferSize)
|
| + ") must be a power of two between 256 and 16384.");
|
| }
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| refNode(node.get()); // context keeps reference until we stop making javascript rendering callbacks
|
| @@ -556,7 +556,7 @@ PassRefPtr<DelayNode> AudioContext::createDelay(double maxDelayTime, ExceptionSt
|
| lazyInitialize();
|
| RefPtr<DelayNode> node = DelayNode::create(this, m_destinationNode->sampleRate(), maxDelayTime, exceptionState);
|
| if (exceptionState.hadException())
|
| - return 0;
|
| + return nullptr;
|
| return node;
|
| }
|
|
|
| @@ -579,7 +579,7 @@ PassRefPtr<ChannelSplitterNode> AudioContext::createChannelSplitter(size_t numbe
|
| "number of outputs (" + String::number(numberOfOutputs)
|
| + ") must be between 1 and "
|
| + String::number(AudioContext::maxNumberOfChannels()) + ".");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| return node;
|
| @@ -604,7 +604,7 @@ PassRefPtr<ChannelMergerNode> AudioContext::createChannelMerger(size_t numberOfI
|
| "number of inputs (" + String::number(numberOfInputs)
|
| + ") must be between 1 and "
|
| + String::number(AudioContext::maxNumberOfChannels()) + ".");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| return node;
|
| @@ -632,14 +632,14 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| exceptionState.throwDOMException(
|
| SyntaxError,
|
| "invalid real array");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| if (!imag) {
|
| exceptionState.throwDOMException(
|
| SyntaxError,
|
| "invalid imaginary array");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| if (real->length() != imag->length()) {
|
| @@ -648,7 +648,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| "length of real array (" + String::number(real->length())
|
| + ") and length of imaginary array (" + String::number(imag->length())
|
| + ") must match.");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| if (real->length() > 4096) {
|
| @@ -656,7 +656,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| IndexSizeError,
|
| "length of real array (" + String::number(real->length())
|
| + ") exceeds allowed maximum of 4096");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| if (imag->length() > 4096) {
|
| @@ -664,7 +664,7 @@ PassRefPtr<PeriodicWave> AudioContext::createPeriodicWave(Float32Array* real, Fl
|
| IndexSizeError,
|
| "length of imaginary array (" + String::number(imag->length())
|
| + ") exceeds allowed maximum of 4096");
|
| - return 0;
|
| + return nullptr;
|
| }
|
|
|
| lazyInitialize();
|
|
|