| Index: third_party/WebKit/LayoutTests/imported/wpt/mediacapture-streams/MediaStreamTrack-MediaElement-disabled-audio-is-silence.html
|
| diff --git a/third_party/WebKit/LayoutTests/imported/wpt/mediacapture-streams/MediaStreamTrack-MediaElement-disabled-audio-is-silence.html b/third_party/WebKit/LayoutTests/imported/wpt/mediacapture-streams/MediaStreamTrack-MediaElement-disabled-audio-is-silence.html
|
| deleted file mode 100644
|
| index 189bb849ee33495fb7e385b641e3b7270c183f07..0000000000000000000000000000000000000000
|
| --- a/third_party/WebKit/LayoutTests/imported/wpt/mediacapture-streams/MediaStreamTrack-MediaElement-disabled-audio-is-silence.html
|
| +++ /dev/null
|
| @@ -1,55 +0,0 @@
|
| -<!doctype html>
|
| -<html>
|
| -<head>
|
| -<title>A disabled audio track is rendered as silence</title>
|
| -<link rel="author" title="Dominique Hazael-Massieux" href="mailto:dom@w3.org"/>
|
| -<link rel="help" href="http://dev.w3.org/2011/webrtc/editor/getusermedia.html#introduction">
|
| -<link rel="help" href="http://dev.w3.org/2011/webrtc/editor/getusermedia.html#mediastreams-as-media-elements">
|
| -</head>
|
| -<body>
|
| -<p class="instructions">When prompted, accept to share your audio stream.</p>
|
| -<h1 class="instructions">Description</h1>
|
| -<p class="instructions">This test checks that a disabled audio track in a
|
| -MediaStream is rendered as silence. It relies on the
|
| -<a href="https://dvcs.w3.org/hg/audio/raw-file/tip/webaudio/specification.html">
|
| -Web Audio API</a>.</p>
|
| -
|
| -<div id='log'></div>
|
| -<script src=/resources/testharness.js></script>
|
| -<script src=/resources/testharnessreport.js></script>
|
| -<script src="/common/vendor-prefix.js" data-prefixed-objects='[{"ancestors":["navigator"], "name":"getUserMedia"}, {"ancestors":["window"], "name":"AudioContext"}]'></script>
|
| -<script>
|
| -var t = async_test("Tests that a disabled audio track in a MediaStream is rendered as silence", {timeout: 200000});
|
| -var aud = document.getElementById("aud");
|
| -t.step(function() {
|
| - navigator.getUserMedia({audio: true}, t.step_func(function (stream) {
|
| - var ctx = new AudioContext();
|
| - var streamSource = ctx.createMediaStreamSource(stream);
|
| - var silenceDetector = ctx.createScriptProcessor(1024);
|
| - var count = 10;
|
| - silenceDetector.onaudioprocess = t.step_func(function (e) {
|
| - var buffer1 = e.inputBuffer.getChannelData(0);
|
| - var buffer2 = e.inputBuffer.getChannelData(1);
|
| - var out = e.outputBuffer.getChannelData(0);
|
| - out = new Float32Array(buffer1);
|
| - for (var i = 0; i < buffer1.length; i++) {
|
| - assert_equals(buffer1[i], 0, "Audio buffer entry #" + i + " in channel 0 is silent");
|
| - }
|
| - for (var i = 0; i < buffer2.length; i++) {
|
| - assert_equals(buffer2[i], 0, "Audio buffer entry #" + i + " in channel 1 is silent");
|
| - }
|
| - count--;
|
| - if (count === 0) {
|
| - silenceDetector.onaudioprocess = null;
|
| - t.done();
|
| - }
|
| - });
|
| - stream.getAudioTracks()[0].enabled = false;
|
| -
|
| - streamSource.connect(silenceDetector);
|
| - silenceDetector.connect(ctx.destination);
|
| - }), function(error) {});
|
| -});
|
| -</script>
|
| -</body>
|
| -</html>
|
|
|