DEV Community

Cover image for Analyzing HLS Audio Streams with Web Audio API and hls.js
Andrey Burov
Andrey Burov

Posted on

Analyzing HLS Audio Streams with Web Audio API and hls.js

When working with Web Audio API and HTML media elements, developers frequently encounter three main issues:

  1. The MediaElementAudioSourceNode Duplication Error

The most common error looks like this:

Cannot create multiple MediaElementAudioSourceNode from the same HTMLMediaElement
Enter fullscreen mode Exit fullscreen mode

This happens because browsers allow creating only one MediaElementAudioSourceNode per <video> or <audio> element. Attempting to create a second one will throw an error, and the original element may lose its audio.

  1. Autoplay Policy Restrictions

Modern browsers block automatic audio playback before user interaction. AudioContext is created in a suspended state, and you must explicitly call resume() after a user gesture (click, tap).

  1. HLS Manifest Loading Synchronization

The HLS manifest loads asynchronously through hls.js. If you create MediaElementAudioSourceNode before the manifest fully loads and attaches to the video element, you'll get silent audio or initialization errors.

The Solution: Singleton Pattern for AudioContext Management

To guarantee a single AudioContext instance and correct MediaElementAudioSourceNode functionality, we use the Singleton pattern. This ensures:

  • Single instance of AudioContext per application
  • Single MediaElementAudioSourceNode per media element
  • Centralized audio graph state management
  • Easy reusability across app components

Basic Implementation

export class HlsAudioService {
  private static instance: HlsAudioService;
  public audioContext: AudioContext;
  public source?: MediaElementAudioSourceNode;
  public splitter?: ChannelSplitterNode;
  public analysers: AnalyserNode[] = [];

  private constructor() {
    const AudioContextClass = window.AudioContext || (window as any).webkitAudioContext;
    this.audioContext = new AudioContextClass();
  }

  static getInstance(): HlsAudioService {
    if (!HlsAudioService.instance) {
      HlsAudioService.instance = new HlsAudioService();
    }
    return HlsAudioService.instance;
  }

  async resumeContext(): Promise<void> {
    if (this.audioContext.state === 'suspended') {
      await this.audioContext.resume();
    }
  }
}

Enter fullscreen mode Exit fullscreen mode

We create a private constructor that initializes AudioContext with cross-browser compatibility (using the webkit prefix for older Safari versions). The static getInstance() method guarantees only one instance throughout the application.

Integration with hls.js

The key is correct initialization sequence. You must wait for the MANIFEST_PARSED event from hls.js before creating audio nodes:

const videoElement = document.querySelector('video') as HTMLVideoElement;
const hls = new Hls();
const audioService = HlsAudioService.getInstance();

hls.on(Hls.Events.MANIFEST_PARSED, () => {
  // Manifest loaded, safe to create audio graph
  if (!audioService.source) {
    audioService.source = audioService.audioContext.createMediaElementSource(videoElement);
    setupAudioGraph(audioService);
  }
});

hls.loadSource('https://example.com/stream.m3u8');
hls.attachMedia(videoElement);

// Handle user gesture
videoElement.addEventListener('play', async () => {
  await audioService.resumeContext();
});
Enter fullscreen mode Exit fullscreen mode

This approach ensures MediaElementAudioSourceNode is created only after hls.js fully initializes the stream, and only once.

Building the Audio Graph for Analysis

After creating the source node, build the processing chain. For stereo analysis, typical architecture looks like:

function setupAudioGraph(service: HlsAudioService) {
  const { audioContext, source } = service;

  // Split stereo signal into left and right channels
  service.splitter = audioContext.createChannelSplitter(2);

  // Create analysers for each channel
  const analyserLeft = audioContext.createAnalyser();
  const analyserRight = audioContext.createAnalyser();

  analyserLeft.fftSize = 2048; // FFT size for frequency analysis
  analyserRight.fftSize = 2048;

  // Build chain: source → splitter → analysers → destination
  source!.connect(service.splitter);
  service.splitter.connect(analyserLeft, 0);  // Left channel
  service.splitter.connect(analyserRight, 1); // Right channel

  // Connect to output for playback
  analyserLeft.connect(audioContext.destination);
  analyserRight.connect(audioContext.destination);

  service.analysers = [analyserLeft, analyserRight];
}
Enter fullscreen mode Exit fullscreen mode

ChannelSplitterNode splits the stereo signal into mono channels. Each AnalyserNode provides frequency and amplitude data for its channel in real-time. The fftSize parameter determines frequency analysis detail — higher values give more resolution but increase CPU load.

Handling Common Edge Cases

CORS and Cross-Origin Streams

If the HLS stream is on another domain, MediaElementAudioSourceNode outputs zeros for security reasons. The solution is adding the crossOrigin attribute to the video element and ensuring the server sends the Access-Control-Allow-Origin header:

videoElement.crossOrigin = 'anonymous';
Enter fullscreen mode Exit fullscreen mode

Without proper CORS setup, audio analysis is impossible, though playback works fine.

Suspended State on Mobile

On mobile platforms (especially iOS), AudioContext may enter suspended state to save battery. Check the state before each use:

async function ensureAudioContextRunning(service: HlsAudioService) {
  if (service.audioContext.state === 'suspended') {
    await service.audioContext.resume();
    console.log('AudioContext resumed');
  }
}
Enter fullscreen mode Exit fullscreen mode

Call this function in play and canplay event handlers.

Switching Between Streams

When changing HLS source (quality switching or channel change), don’t recreate MediaElementAudioSourceNode. Simply call hls.loadSource() with the new URL — the existing audio graph continues working:

function switchStream(newUrl: string) {
  hls.loadSource(newUrl);
  // source node remains, recreation not needed
}
Enter fullscreen mode Exit fullscreen mode

Attempting to recreate the source will error since it’s already bound to the element.

Performance Optimization

To reduce CPU load during visualization:

Reduce update frequency: Instead of updating every frame (60 FPS), limit to 30 FPS using requestAnimationFrame and frame counting.

Adjust FFT size: Simple level indicators need fftSize = 256, detailed spectrograms need 2048 or 4096. Smaller values reduce latency and load.

Web Workers: You can offload data processing from AnalyserNode to a worker, avoiding main-thread blocking. However, audio nodes themselves must stay on the main thread.

Use Cases

With the obtained data, you can implement various visualizations and analyses:

  • PPM/VU meters — display current signal level with different integration times
  • Spectrograms — real-time frequency spectrum visualization using getByteFrequencyData()
  • Silence detector — analyze amplitude to detect audio pauses
  • Equalizers — divide frequencies into bands for volume control

All these tasks use data from AnalyserNode, which provides an array of values from 0 to 255 for each frequency band.

Browser Compatibility

Browser Web Audio API Native HLS HLS.js (MSE)
Chrome (Desktop) v14+ No Yes
Firefox (Desktop) v25+ No Yes
Safari (Desktop) v6.1+ Yes ⚠Not needed
Edge Full No Yes
Chrome Mobile Full Yes (Android) Yes
Safari iOS v6+ Yes No (MSE)
Firefox Android Full No Yes

Important: Safari on iOS doesn’t support Media Source Extensions, so hls.js won’t work. Fortunately, iOS has native HLS support, and you can use video.src directly. Web Audio API works correctly with it.

Common Errors Reference

Error Cause Solution
Cannot create multiple MediaElementAudioSourceNode Attempting to create MediaElementAudioSourceNode twice for same element Use Singleton pattern to maintain single reference
AudioContext was not allowed to start AudioContext created before user interaction Call audioContext.resume() after user gesture (click, touch)
MediaElementAudioSourceNode outputs zeroes CORS restrictions for cross-domain audio Add crossOrigin="anonymous" to video element
HLS manifest not loading HLS manifest loads before AudioContext initialization Wait for MANIFEST_PARSED event before creating source node

Conclusion

Analyzing audio from HLS streams in the browser is achievable with the right approach. Key takeaways:

  1. Singleton for AudioContext prevents node recreation errors and simplifies state management
  2. Waiting for MANIFEST_PARSED from hls.js guarantees correct initialization
  3. Handling autoplay policy via resume() ensures cross-platform compatibility
  4. CORS setup is essential for cross-domain streams

A complete working implementation is available at github.com/ABurov30/AudioContext, where you can study ready-made integration of all described techniques.

This approach enables building reliable web applications for streaming video with advanced audio analysis, without requiring users to install additional plugins or extensions.

Top comments (0)