DXR is a code search and navigation tool aimed at making sense of large projects. It supports full-text and regex searches as well as structural queries.

Line Code
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101
<!DOCTYPE HTML>
<html>
<head>
  <title>Test AudioCapture </title>
  <script type="application/javascript" src="mediaStreamPlayback.js"></script>
</head>
<body>
<pre id="test">
<script>

createHTML({
  bug: "1156472",
  title: "Test AudioCapture with regular HTMLMediaElement, AudioContext, and HTMLMediaElement playing a MediaStream",
  visible: true
});

// Get an opus file containing a sine wave at maximum amplitude, of duration
// `lengthSeconds`, and of frequency `frequency`.
async function getSineWaveFile(frequency, lengthSeconds) {
  const off = new OfflineAudioContext(1, lengthSeconds * 48000, 48000);
  const osc = off.createOscillator();
  const rec = new MediaRecorder(osc, {mimeType: "audio/ogg; codecs=opus"});
  osc.frequency.value = frequency;
  osc.start();
  rec.start();
  off.startRendering();
  const {data} = await new Promise(r => rec.ondataavailable = r);
  return data;
}

(async function() {
  await scriptsReady;
  FAKE_ENABLED = false;
  await runTestWhenReady(async function() {
    /**
     * Get two HTMLMediaElements:
     * - One playing a sine tone from a blob (of an opus file created on the fly)
     * - One being the output for an AudioContext's OscillatorNode, connected to
     *   a MediaSourceDestinationNode.
     *
     * Also, use the AudioContext playing through its AudioDestinationNode another
     * tone, using another OscillatorNode.
     *
     * Capture the output of the document, feed that back into the AudioContext,
     * with an AnalyserNode, and check the frequency content to make sure we
     * have recorded the three sources.
     *
     * The three sine tones have frequencies far apart from each other, so that we
     * can check that the spectrum of the capture stream contains three
     * components with a high magnitude.
     */
    const wavtone = createMediaElement("audio", "WaveTone");
    const acTone = createMediaElement("audio", "audioContextTone");
    const ac = new AudioContext();

    const oscThroughMediaElement = ac.createOscillator();
    oscThroughMediaElement.frequency.value = 1000;
    const oscThroughAudioDestinationNode = ac.createOscillator();
    oscThroughAudioDestinationNode.frequency.value = 5000;
    const msDest = ac.createMediaStreamDestination();

    oscThroughMediaElement.connect(msDest);
    oscThroughAudioDestinationNode.connect(ac.destination);

    acTone.srcObject = msDest.stream;

    const blob = await getSineWaveFile(10000, 10);
    wavtone.src = URL.createObjectURL(blob);
    oscThroughMediaElement.start();
    oscThroughAudioDestinationNode.start();
    wavtone.loop = true;
    wavtone.play();
    acTone.play();

    const constraints = {audio: {mediaSource: "audioCapture"}};

    const stream = await getUserMedia(constraints);
    window.grip = stream;
    const analyser = new AudioStreamAnalyser(ac, stream);
    analyser.enableDebugCanvas();
    await analyser.waitForAnalysisSuccess(array => {
      // We want to find three frequency components here, around 1000, 5000
      // and 10000Hz. Frequency are logarithmic. Also make sure we have low
      // energy in between, not just a flat white noise.
      return (array[analyser.binIndexForFrequency(50)]    < 50 &&
              array[analyser.binIndexForFrequency(1000)]  > 200 &&
              array[analyser.binIndexForFrequency(2500)]  < 50 &&
              array[analyser.binIndexForFrequency(5000)]  > 200 &&
              array[analyser.binIndexForFrequency(7500)]  < 50 &&
              array[analyser.binIndexForFrequency(10000)] > 200);
    });
    finish();
  });
})();



</script>
</pre>
</body>
</html>