| Index: Source/modules/webaudio/AudioContext.h
|
| diff --git a/Source/modules/webaudio/AudioContext.h b/Source/modules/webaudio/AudioContext.h
|
| index 021168019b2129b36a502dc25768272d78bd85e7..fde6ef7ed7740ff95a48abdf0ce71b0919cf28bc 100644
|
| --- a/Source/modules/webaudio/AudioContext.h
|
| +++ b/Source/modules/webaudio/AudioContext.h
|
| @@ -25,8 +25,6 @@
|
| #ifndef AudioContext_h
|
| #define AudioContext_h
|
|
|
| -#include "bindings/core/v8/ScriptPromise.h"
|
| -#include "bindings/core/v8/ScriptPromiseResolver.h"
|
| #include "core/dom/ActiveDOMObject.h"
|
| #include "core/events/EventListener.h"
|
| #include "modules/EventTargetModules.h"
|
| @@ -79,16 +77,6 @@
|
| DEFINE_WRAPPERTYPEINFO();
|
| WILL_BE_USING_GARBAGE_COLLECTED_MIXIN(AudioContext);
|
| public:
|
| - // The state of an audio context. On creation, the state is Paused. The state is Running if
|
| - // audio is being processed (audio graph is being pulled for data). The state is Released if the
|
| - // audio context has been released. The valid transitions are from Paused to either Running or
|
| - // Released; Running to Paused or Released. Once Released, there are no valid transitions.
|
| - enum AudioContextState {
|
| - Paused,
|
| - Running,
|
| - Released
|
| - };
|
| -
|
| // Create an AudioContext for rendering to the audio hardware.
|
| static AudioContext* create(Document&, ExceptionState&);
|
|
|
| @@ -107,7 +95,6 @@
|
| size_t currentSampleFrame() const { return m_destinationNode->currentSampleFrame(); }
|
| double currentTime() const { return m_destinationNode->currentTime(); }
|
| float sampleRate() const { return m_destinationNode->sampleRate(); }
|
| - String state() const;
|
|
|
| AudioBuffer* createBuffer(unsigned numberOfChannels, size_t numberOfFrames, float sampleRate, ExceptionState&);
|
|
|
| @@ -141,10 +128,6 @@
|
| OscillatorNode* createOscillator();
|
| PeriodicWave* createPeriodicWave(Float32Array* real, Float32Array* imag, ExceptionState&);
|
|
|
| - // Pause/Resume
|
| - void suspendContext(ExceptionState&);
|
| - ScriptPromise resumeContext(ScriptState*);
|
| -
|
| // When a source node has no more processing to do (has finished playing), then it tells the context to dereference it.
|
| void notifyNodeFinishedProcessing(AudioNode*);
|
|
|
| @@ -291,22 +274,6 @@
|
| // AudioNode::breakConnection() when we remove an AudioNode from this.
|
| HeapVector<Member<AudioNode> > m_referencedNodes;
|
|
|
| - // Stop rendering the audio graph.
|
| - void stopRendering();
|
| -
|
| - // Handle Promises for resume().
|
| - void resolvePromisesForResume();
|
| - void resolvePromisesForResumeOnMainThread();
|
| -
|
| - // Vector of promises created by resume(). It takes time to handle them, so we collect all of
|
| - // the promises here until they can be resolved or rejected.
|
| - Vector<RefPtr<ScriptPromiseResolver> > m_resumePromises;
|
| -
|
| - // True if we're in the process of resolving promises for resume(). Resolving can take some
|
| - // time and the audio context process loop is very fast, so we don't want to call resolve an
|
| - // excessive number of times.
|
| - bool m_isResolvingResumePromises;
|
| -
|
| class AudioNodeDisposer {
|
| public:
|
| explicit AudioNodeDisposer(AudioNode& node) : m_node(node) { }
|
| @@ -371,9 +338,6 @@
|
|
|
| bool m_isOfflineContext;
|
|
|
| - AudioContextState m_contextState;
|
| - void setContextState(AudioContextState);
|
| -
|
| AsyncAudioDecoder m_audioDecoder;
|
|
|
| // Collection of nodes where the channel count mode has changed. We want the channel count mode
|
|
|