The WebRTC market is experiencing explosive growth in 2026. According to Technavio, the market is projected to expand by USD 247.7 billion from 2025 to 2029, representing a staggering 62.6% compound annual growth rate. These aren't just incremental shifts—the WebRTC trends in 2026 represent a fundamental transformation of how real-time communication infrastructure works at scale.
WebRTC (Web Real-Time Communication) enables peer-to-peer audio, video, and data sharing directly in web browsers without plugins or native apps. It's the invisible infrastructure powering video calls, live streaming, telehealth consultations, and collaborative tools used by billions of people daily. At the core of reliable WebRTC connectivity is a TURN server—the relay that ensures connections work even behind restrictive NATs and firewalls.
Why 2025 was a pivotal year? Three forces are converging: AI integration is moving from experimental to production, new protocols like Media over QUIC are reshaping streaming architecture, and market adoption is accelerating across industries from telehealth to IoT.
Here are the 7 trends defining WebRTC in 2026:
- AI & Machine Learning Integration — Real-time translation, noise suppression, and voice agents
- Media over QUIC (MoQ) Protocol Emergence — Combining WebRTC latency with broadcast scale
- Codec Evolution — AV1, VP9, and H.265 bandwidth optimization
- IoT & Edge Computing — 18 billion devices by year-end
- AR/VR/XR Expansion — Spatial audio and cross-platform immersive experiences
- Security & Privacy Enhancements — DTLS 1.3 migration and SFrame E2EE
- Market Growth & Industry Adoption — Telehealth, enterprise, and SME acceleration
From an infrastructure operator's perspective, these trends have profound implications for TURN relay architecture, bandwidth economics, and global connectivity. Let's explore what's really happening beneath the surface.
Trend 1 — AI & Machine Learning Integration: The Dominant Force
AI integration isn't just a trend—it's reshaping the entire WebRTC landscape. By 2024, WebRTC already underpinned 89% of real-time internet communication, and the market is projected to surge from $19.4 billion in 2025 to $755.5 billion by 2035, driven primarily by AI applications.
But here's what most coverage misses: the infrastructure requirements are fundamentally different.
OpenAI Realtime API and WebRTC
In December 2024, OpenAI announced WebRTC Endpoint support for their Realtime API. This closed a critical gap for integrating large language models with real-time voice communication. Now developers can build AI voice agents that respond to users through WebRTC connections with minimal latency.
The use cases are already emerging. Conversational AI assistants that handle customer service calls in real-time. Voice-first applications where users speak naturally to AI systems. Interactive tutoring platforms where AI responds instantly to student questions.
Here's the catch: AI voice agents demand sub-300ms end-to-end latency for natural conversation. That's significantly stricter than typical WebRTC video calls, where 500-800ms is often acceptable. When you're talking to an AI, every 100ms of additional delay breaks the illusion of natural interaction.
Practical AI Applications in WebRTC
AI is enhancing WebRTC in ways that were science fiction just two years ago.
Real-time translation now works during live video calls. Machine learning models automatically translate spoken language as people speak, enabling seamless multilingual conversations. Japanese and English speakers can collaborate in real-time without either learning the other's language.
Noise suppression has evolved beyond simple filters. ML models isolate human voices from ambient noise—barking dogs, construction sounds, keyboard typing—and suppress them in real-time without degrading voice quality. The model learns what's "voice" and what's "noise" and adapts continuously.
Video upscaling improves low-resolution streams on the fly. When someone joins from a poor connection or older device, AI models enhance the video quality dynamically, adjusting compression based on content complexity. A static talking head gets more compression than a screen share with detailed text.
Sentiment analysis is being deployed in customer service applications. The system gauges emotions through tone, pitch, and content, alerting human agents when users become frustrated. This allows preemptive intervention before customers churn.
Sign language translation represents a breakthrough for accessibility. Real-time computer vision models can interpret sign language and convert it to speech or text, enabling deaf and hard-of-hearing users to participate in voice calls without human interpreters.
Technical Implementation
How does this actually work? TensorFlow.js enables developers to run machine learning models directly in web browsers. This means AI processing can happen client-side without round-tripping to a server, reducing latency and protecting privacy.
Edge AI integration is accelerating this trend. Instead of centralizing all processing in the cloud, computation happens at the network edge—closer to users. This decentralizes the load, reduces latency, and improves reliability when cloud connectivity is intermittent.
The architecture looks like this: browser captures audio/video → TensorFlow.js model processes locally → enhanced stream sent over WebRTC → recipient receives improved quality. All in real-time, all while maintaining sub-300ms latency.
Infrastructure Implications: The Hidden Challenge
Here's what the AI hype doesn't mention: global TURN relay architecture becomes critical when you need <300ms latency.
Consider the scenario: A user in Singapore talks to an AI voice agent hosted in US-East. The round-trip network latency alone—Singapore to Virginia and back—is roughly 200-250ms under ideal conditions. Add encoding, decoding, and processing time, and you're already approaching or exceeding the 300ms budget.
The solution? Global TURN relay with optimized routing. When the user in Singapore connects through a local TURN server, and that TURN server has a private, high-speed connection to the region hosting the AI, you can shave 50-100ms off the total latency. That's the difference between natural conversation and noticeable lag.
AI voice agents also create different traffic patterns than traditional peer-to-peer WebRTC. Instead of bursty video calls that last 20-40 minutes, AI applications often involve sustained connections with unpredictable spikes. A customer service AI might handle hundreds of simultaneous conversations, each requiring low-latency relay.
Bandwidth considerations matter too. While the audio itself is lightweight (typically 32-64 kbps), AI-enhanced video with real-time upscaling can demand 2-3x typical bitrates during processing. Infrastructure needs to handle these bursts without degrading quality.
The economics are shifting as well. Traditional WebRTC operates on a peer-to-peer model where TURN relay is only needed when direct connection fails (roughly 15-20% of cases). AI voice agents always go through infrastructure—there is no peer-to-peer fallback. This means 100% of traffic hits TURN servers, fundamentally changing cost modeling and capacity planning.
Trend 2 — Media over QUIC (MoQ): Protocol Evolution
A new protocol is emerging that could reshape streaming architecture. Media over QUIC (MoQ) combines the low latency of WebRTC with the scale of traditional streaming protocols like HLS and DASH, all while simplifying the technical complexity that has plagued real-time streaming for years.
But before you rip out your WebRTC infrastructure, here's the reality check: MoQ is promising, but production readiness is still 2026+.
What is Media over QUIC?
MoQ is an open protocol being developed at the IETF by engineers from Google, Meta, Cisco, Akamai. The goal is ambitious: solve what's been called the "historical trilemma" of streaming.
For decades, you could have two of these three, but not all three:
- Sub-second latency (like WebRTC)
- Broadcast scale (like HLS/DASH serving millions of viewers)
- Architectural simplicity (not requiring complex server-side processing)
Traditional WebRTC gives you low latency but struggles at broadcast scale—sending 1080p video to 100,000 viewers simultaneously is expensive and complex. HLS/DASH scales beautifully to millions of viewers but has 10-30 seconds of latency. RTMP was simple but had neither scale nor latency.
MoQ aims to deliver all three by treating media as subscribable tracks in a publish/subscribe system designed specifically for real-time media at CDN scale. Instead of point-to-point connections, media flows through relay entities that can cache, forward, and distribute efficiently.
MoQ vs WebRTC — Complementary, Not Competitive
Here's a key insight that gets missed in breathless coverage: MoQ and WebRTC are complementary technologies, not competitors.
WebRTC excels at interactive, bidirectional communication. Think video conferencing where everyone can talk, screen sharing in collaborative tools, or peer-to-peer file transfers. The interactivity is the point—low latency matters because participants need to respond to each other in real-time.
MoQ is designed for scalable, broadcast-scale streaming with sub-second latency. Think live sports streaming to millions, concert broadcasts where viewers don't need to talk back, or large-scale webinars where one presenter addresses thousands. The distribution is the point—reaching massive audiences while maintaining live-like latency.
The decision framework is straightforward:
- Use WebRTC when: You need bidirectional communication, fewer than 100 participants, or interactive features like screen sharing
- Use MoQ when: You need to stream to thousands or millions, viewers don't need to send media back, or you want CDN-friendly distribution
Some applications will use both. A large webinar might use MoQ to broadcast the presenter to 10,000 viewers, while using WebRTC for the Q&A panel of 5-10 speakers who need to interact.
Production Status & Browser Support: The 2026 Reality Check
But here's where we need to be cautiously optimistic rather than prematurely enthusiastic.
Browser support is incomplete. Chrome and Edge (Chromium-based browsers) support WebTransport, which MoQ relies on. Safari doesn't yet have fully functional WebTransport support, though Apple has indicated their intent to implement it. Until Safari supports it, you're cutting off a significant chunk of mobile and desktop users.
Production readiness is still developing. As of December 2024, industry consensus is that MoQ isn't quite ready for production use cases, though it's coming soon given current momentum. Red5, a major streaming platform vendor, plans to support MoQ by the end of 2025—that's a concrete timeline indicating when production deployment becomes realistic.
The workhorses are still VP8 and H.264. For all the excitement around new protocols, the vast majority of WebRTC traffic in 2025 runs on battle-tested codecs and proven architectures. MoQ represents the future, but that future is 2026 and beyond, not today.
This doesn't mean ignore MoQ. It means watch this space, understand the architecture, and prepare your infrastructure to adapt when adoption reaches critical mass. Early movers who understand MoQ will have competitive advantages when it matures.
Infrastructure Implications: How TURN Adapts
What does MoQ mean for TURN relay infrastructure? The architecture is different but the need for relay doesn't disappear—it transforms.
MoQ introduces relay entities that forward media over QUIC or HTTP/3. These aren't traditional TURN servers, but they serve a similar function: relaying media when direct delivery isn't optimal. The key difference is that MoQ relays are designed to work seamlessly with CDNs, allowing existing CDN infrastructure to be upgraded rather than replaced.
For infrastructure operators, this means planning for dual-protocol support. WebRTC TURN servers for interactive use cases will coexist with MoQ relay entities for broadcast scenarios. The two protocols handle different problems, so the infrastructure to support both will be necessary.
The cost model shifts slightly. MoQ's CDN-friendly design means caching becomes possible—the same media stream can be cached at edge locations and delivered to multiple viewers from cache. Traditional TURN relay doesn't allow caching because every connection is unique. This could reduce bandwidth costs for broadcast scenarios while maintaining low latency.
Geographic distribution remains critical. Just like WebRTC benefits from global TURN relay, MoQ will benefit from globally distributed relay entities. Users in APAC shouldn't have to pull streams from US-East—they should hit a local relay that caches or forwards efficiently.
The timeline for infrastructure adaptation is 2026+. Operators can monitor MoQ development, test implementations as they mature, and plan for gradual integration. The transition will be evolutionary, not revolutionary—WebRTC isn't going anywhere, and MoQ will supplement rather than replace it for the foreseeable future.
Trend 3 — Codec Evolution: AV1, VP9, and the Reality Check
Video codecs determine how much bandwidth real-time communication consumes. In 2025, a new generation of codecs promises massive bandwidth savings—but the reality is more nuanced than the hype suggests.
AV1 — Promise vs Reality
AV1 is the darling of codec discussions. Developed by the Alliance for Open Media (a consortium including Google, Mozilla, Cisco, and others), AV1 is royalty-free and delivers impressive compression efficiency. At equivalent video quality, AV1 reduces file sizes by 30-50% compared to VP9 and H.265.
The bandwidth savings are real. Testing shows AV1 performs exceptionally well at low bitrates—200 to 600 kbps—maintaining excellent visual quality even under constrained bandwidth conditions. For users on mobile networks or in regions with poor connectivity, this is transformative.
Here's the reality check: AV1 encoding is 5 to 10 times slower than VP9, and CPU usage can peak at 225% during active encoding. That's not a typo—it's more than double the CPU load compared to VP9.
For live, real-time applications like video conferencing, this matters enormously. You can't pre-encode AV1 content in advance like you can for video-on-demand. The encoding must happen in real-time as users speak, and if your device can't keep up, the stream degrades or drops frames.
Hardware acceleration is improving. Newer GPUs and dedicated encoding chips are adding AV1 support, which brings CPU usage down to manageable levels. But hardware support isn't universal yet—especially on mobile devices and older laptops that are still widely used in 2025.
The practical takeaway? AV1 is coming, but it's not the default for real-time WebRTC in 2025. It's being adopted gradually, particularly in scenarios where users have modern hardware and bandwidth is constrained. Think mobile networks in developing markets, or high-quality screen sharing where text clarity matters more than smooth motion.
VP9 — The Workhorse
While everyone talks about AV1, VP9 quietly powers the majority of high-quality WebRTC streams in 2025. Why? It strikes the best balance between compression efficiency, CPU usage, and feature support.
VP9 is the only codec in WebRTC that supports Scalable Video Coding (SVC). SVC allows a single video stream to be encoded at multiple quality levels simultaneously, and recipients can subscribe to the layer that matches their bandwidth and device capabilities.
This is critical for large group video calls and live broadcasts. Instead of encoding three separate streams (high, medium, low quality), you encode once with SVC, and the server forwards the appropriate layer to each participant. It's vastly more efficient for group scenarios.
VP9 also has mature hardware support across devices. Nearly all modern smartphones, laptops, and browsers can encode and decode VP9 efficiently. The ecosystem is battle-tested and stable.
For most WebRTC deployments in 2025, VP9 remains the ideal choice for group calls, webinars, and any scenario requiring SVC. The compression is good (not quite as good as AV1, but close), CPU usage is reasonable, and it just works reliably across the ecosystem.
H.265 (HEVC) — The Enterprise Option
H.265 (also known as HEVC) is an interesting middle ground. It offers strong compression efficiency—close to VP9—and has excellent hardware encoder support, resulting in low CPU usage on supported devices.
Chrome 136 Beta added H.265 hardware encoder support, signaling broader adoption. When hardware acceleration is available, H.265 can deliver high-quality video with minimal CPU load, making it attractive for enterprise deployments where devices are newer and more powerful.
The challenge? H.265 has limited WebRTC and browser support due to licensing issues. Patent licensing fees make it economically complicated for open-source projects and free-tier services. Apple devices support it well, but broad cross-platform support lags behind royalty-free alternatives like VP8, VP9, and AV1.
For enterprise use cases where all participants are on managed devices with H.265 support, it's a viable option. For general-purpose web applications reaching diverse audiences, VP9 or VP8 remains safer.
Codec Selection Decision Framework
Here's how to choose:
| Codec | When to Use | Best For | Limitations |
|---|---|---|---|
| AV1 | Bandwidth-constrained environments, modern hardware with acceleration | Mobile networks, low-bandwidth scenarios, screen sharing with text | High CPU usage without hardware support; encoding 5-10× slower than VP9 |
| VP9 | Group calls, webinars, broadcasts requiring SVC | Large meetings (10+ participants), live streaming to multiple bitrates | Slightly higher bandwidth than AV1; less hardware support than H.264 |
| H.264 | Maximum compatibility, legacy device support | Public-facing applications, broad audience reach | Larger file sizes; older compression technology |
| H.265 | Enterprise deployments with known hardware, low CPU budget | Managed corporate environments, Apple ecosystem | Limited browser support due to licensing; not universal |
The reality for 2025: VP8 and H.264 remain the workhorses for most WebRTC services. VP9 is the go-to for SVC use cases. AV1 is being adopted gradually as hardware support expands. H.265 serves niche enterprise scenarios.
Infrastructure Implications: Bandwidth Economics
From an infrastructure operator's perspective, codec evolution directly impacts bandwidth costs and relay performance.
AV1 adoption means 30-50% bandwidth savings when it reaches scale. For a TURN relay provider handling petabytes of traffic monthly, that translates to significant cost reduction—potentially millions of dollars annually at large scale. But the transition won't happen overnight.
The CPU vs bandwidth trade-off is real. Operators must decide whether to push encoding to clients (saving relay server CPU but requiring capable client devices) or handle transcoding server-side (consuming server CPU but supporting any client). This affects hardware procurement, power consumption, and operational costs.
Codec negotiation complexity increases. Supporting multiple codecs means relay infrastructure must handle fallback scenarios gracefully. When a VP9-capable sender connects to an H.264-only recipient, who transcodes? Where does it happen? These architectural decisions cascade through infrastructure design.
Relay performance varies by codec. Some codecs handle packet loss better than others. AV1's advanced error resilience means it degrades more gracefully when network conditions deteriorate. Infrastructure operators can optimize retry logic and forward error correction based on which codecs are in use.
The long-term outlook is clear: gradual AV1 adoption through 2025-2026, with VP9 and H.264 maintaining significant market share for years. Infrastructure must support all of them simultaneously, optimizing for the codecs that see the most traffic while preparing for the shift toward next-generation compression.
Trend 4 — IoT & Edge Computing: 18 Billion Devices by Year-End
The Internet of Things is exploding, and WebRTC is becoming the communication protocol of choice for real-time IoT applications. By the end of 2025, an estimated 18 billion IoT devices will be online worldwide, generating a staggering 79.4 zettabytes of data according to IDC.
Most people associate WebRTC with video calls, but IoT represents a fundamentally different use case—and one that's growing faster than anyone predicted.
IoT Device Explosion
The types of devices adopting WebRTC might surprise you. We're not just talking about smart displays or video doorbells (though those are significant). The technology is spreading to smoke detectors, thermostats, industrial sensors, and even agricultural equipment.
Smart cameras and video doorbells are the most visible examples. Brands like Ring, Nest, and Arlo use WebRTC to stream real-time video from cameras to smartphones without requiring proprietary apps or cloud relay services (though many still use cloud relay for broader compatibility).
Home automation devices are integrating WebRTC for remote monitoring and control. A thermostat that can stream live video of the room it's in. A smoke detector that can establish a video call to emergency services automatically when triggered.
Industrial IoT is where things get interesting. Factory sensors that stream real-time telemetry and video to remote monitoring centers. Construction site cameras that provide live feeds to project managers without on-site IT infrastructure. Agricultural drones that transmit real-time video during automated inspections.
The common thread? These devices need real-time communication without proprietary apps, cloud dependency, or complex setup. WebRTC provides exactly that—standardized, peer-to-peer (or relay-assisted) communication that works across platforms.
WebRTC in IoT
In 2024, AWS released a WebRTC SDK for Kinesis Video Streams specifically to accelerate smart camera integrations. This makes it dramatically easier for device manufacturers to add WebRTC support without building the entire stack from scratch.
The value proposition is compelling: devices communicate using the same protocol that's already in every web browser. No need for users to install native apps. No need for device manufacturers to maintain separate app codebases for iOS and Android. Just point a browser at a URL, and you're connected to the device.
Edge computing integration is the force multiplier. Instead of sending raw sensor data to the cloud for processing (which consumes bandwidth and adds latency), devices process data locally at the edge. Then they send only the relevant insights or compressed summaries over WebRTC.
Consider a security camera with edge AI. It processes video locally to detect motion or recognize faces. When something interesting happens, it establishes a WebRTC connection to send a real-time alert with the relevant video clip. The bulk of the video never leaves the device—only the important moments get transmitted.
This architecture is more privacy-preserving (raw video doesn't go to the cloud), more bandwidth-efficient (only alerts and clips are sent), and more responsive (detection happens locally without round-trip latency).
Infrastructure Implications: TURN for IoT
Here's the infrastructure challenge that IoT creates: many IoT devices sit behind carrier-grade NAT (CGN) or symmetric NAT, making direct peer-to-peer WebRTC connections impossible.
In residential broadband, users typically get a public IP address (or at least a NAT-friendly configuration). IoT devices often connect via cellular networks where CGN is universal. An LTE-connected security camera might have an internal IP like 100.64.0.5—completely unreachable from the public internet.
The solution? Always-on TURN relay. Unlike typical WebRTC video calls where TURN is a fallback (needed 15-20% of the time), IoT devices behind CGN require TURN 100% of the time. There is no peer-to-peer fallback—the relay is mandatory.
This changes cost modeling fundamentally. If you're deploying 1,000 IoT cameras, you're not planning for 150-200 to use TURN relay. You're planning for all 1,000 to use relay, all the time.
Scaling economics shift accordingly. 18 billion IoT devices by end of 2025 means exponential TURN relay demand. Even if only 1% of those devices use WebRTC for video streaming, that's 180 million devices requiring always-on relay infrastructure. The bandwidth and server capacity implications are massive.
Regional distribution becomes critical. A smart camera in Tokyo shouldn't relay through a TURN server in Virginia. The latency would make real-time monitoring unusable. IoT deployments need geographically distributed TURN infrastructure—APAC, EMEA, North America, Latin America—to provide acceptable latency for global device fleets.
APAC is seeing the fastest growth in IoT adoption, driven by rapid digitalization in India, Southeast Asia, and expanding 5G networks in China and South Korea. Infrastructure operators without strong APAC presence will struggle to serve this market effectively.
Metered's 31+ regions across 5 continents provide the geographic coverage IoT deployments need. When a manufacturer ships cameras to 20 countries, they need relay infrastructure in all 20 countries—not a single region that forces all traffic through intercontinental backhaul.
The opportunity is enormous, but so are the infrastructure demands. IoT isn't just another WebRTC use case—it's a category that dwarfs traditional video conferencing in scale and requires fundamentally different architectural assumptions.
Trend 5 — AR/VR/XR: Immersive Experiences Go Mainstream
Augmented reality, virtual reality, and extended reality (collectively XR) are transitioning from experimental novelty to practical mainstream applications in 2025. WebRTC is the invisible infrastructure making it possible.
XR Market Maturity in 2025
The XR market in 2025 is defined by three factors: the mainstream rise of smart glasses, deeper AI integration, and rapid improvements in display technology.
Smart glasses are going consumer. Meta's Ray-Ban Smart Glasses have signaled growing demand for stylish, functional wearables that blend digital and physical worlds. These aren't the bulky headsets of previous generations—they're glasses that look relatively normal while adding computational layers to what you see.
AI is making XR more intuitive. Real-time object recognition allows glasses to identify objects and provide contextual information. Gesture control eliminates the need for handheld controllers. Generative content means XR environments can adapt dynamically based on what users do.
5G-Advanced is rolling out in 2025, addressing the latency and bandwidth bottlenecks that previously limited XR applications. Lower latency (sub-10ms in ideal conditions) and more reliable connections make it feasible to stream high-fidelity XR content without requiring powerful local hardware.
The convergence of these trends is making XR practical for real use cases: virtual collaboration spaces where distributed teams feel like they're in the same room, immersive training simulations for medical and industrial applications, and entertainment experiences that blend physical and digital worlds.
WebRTC's Role in the Metaverse
Here's something critical that often gets overlooked: WebRTC is currently the only option for transmitting real-time video directly from an AR/VR device to a web browser without requiring plugins or native applications.
Think about the implications. A doctor wearing AR glasses during surgery can stream their point-of-view to a specialist consultant on the other side of the world, who views it in a standard web browser. No app installation required, no complex setup—just a WebRTC connection providing real-time, low-latency video.
Multi-party VR experiences depend on the lowest possible latency to maintain immersion. When you're in a virtual meeting room with colleagues represented as avatars, every millisecond of delay breaks the sense of presence. Voice needs to be synchronized with lip movements and gestures. If someone reaches to shake your (virtual) hand, the delay between their action and your perception can't exceed 50ms or the illusion shatters.
Cross-platform communication is where WebRTC becomes indispensable. Apple Vision Pro users need to communicate with Meta Quest users, who need to communicate with people on flat screens. WebRTC provides the standardized protocol that makes cross-platform XR collaboration possible without each vendor implementing proprietary systems.
Spatial Audio & Advanced Technologies
6-DOF (six degrees of freedom) audio rendering lets listeners move freely in a virtual environment—forward, backward, up, down, left, right—and audio positioning stays consistent with their perspective. When you walk around a virtual speaker, the sound appears to come from the correct direction relative to your position.
This is essential for VR. Without spatial audio, virtual environments feel flat and unconvincing. With it, presence and immersion skyrocket. Dolby has been using WebRTC to improve spatial audio quality, paying particular attention to overlapping speech, laughter, and other aspects of natural communication that previous systems struggled with.
Volumetric video captures people in three dimensions, allowing you to see them from any angle in VR. Instead of a flat video screen floating in virtual space, you see a 3D representation of the person that you can walk around. This is bandwidth-intensive—volumetric video can require 10-50× more bandwidth than traditional 2D video—but the immersion improvement is transformative.
Avatar mirroring uses computer vision to track facial expressions and body language, translating them to virtual avatars in real-time. When you smile, your avatar smiles. When you gesture, your avatar gestures. This maintains non-verbal communication cues that are crucial for natural interaction.
Infrastructure Implications: Ultra-Low Latency Requirements
From an infrastructure perspective, AR/VR applications impose some of the strictest requirements in all of WebRTC.
Latency budgets are brutal. For truly immersive experiences, motion-to-photon latency (the time between head movement and updated visual display) must be under 20ms to prevent motion sickness. Audio-visual synchronization must stay within 50ms to avoid perceptible mismatch. End-to-end network latency needs to be under 50ms for multi-party VR to feel natural.
These aren't aspirational targets—they're hard requirements. Exceed them and users experience discomfort, nausea, or break the sense of presence that makes XR compelling.
Volumetric video bandwidth demands are enormous. While traditional 1080p video might consume 2-4 Mbps, volumetric video can require 20-100 Mbps depending on quality and compression. TURN relay infrastructure must handle these sustained high-bandwidth streams without introducing additional latency or packet loss.
Global relay for cross-continent XR collaboration is where private TURN backbones become critical. Imagine a virtual design review with participants in London, Tokyo, and San Francisco. If each participant routes through their nearest TURN server, and those TURN servers relay media over the public internet, latency will be 200-400ms—unacceptable for immersive collaboration.
Regional distribution matters tremendously. An XR application serving users in Southeast Asia needs TURN servers in Singapore, not just Virginia or Frankfurt. The round-trip latency penalty for forcing APAC traffic through Europe or North America makes immersive experiences impossible.
The opportunity in XR is massive, but the infrastructure demands are unforgiving. Low latency isn't negotiable—it's the difference between an application that works and one that makes users nauseous. Operators who can deliver consistent sub-50ms latency globally will have a decisive advantage as XR goes mainstream.
Trend 6 — Security & Privacy: DTLS 1.3 and SFrame E2EE
WebRTC has mandatory encryption on all components—video, audio, and data channels are always encrypted. But in 2025, the security landscape is evolving with protocol updates and new encryption schemes that respond to emerging threats and regulatory requirements.
DTLS 1.3 Migration (February 2025)
As of February 2025, the WebRTC ecosystem began migrating to DTLS 1.3. Modern browsers are phasing out older ciphers and requiring applications to implement minimum-version negotiation. DTLS 1.0 and 1.1 are being deprecated.
Why does this matter? DTLS (Datagram Transport Layer Security) is the protocol that encrypts WebRTC data channels. The upgrade to 1.3 brings stronger cryptographic primitives, improved performance (reduced handshake round-trips), and removes legacy ciphers that have known vulnerabilities.
For developers, this means updating WebRTC implementations to support DTLS 1.3. For end users, it means stronger security by default with no action required.
SFrame End-to-End Encryption
SFrame is being standardized through the IETF and major WebRTC platforms are adopting it for end-to-end encryption in group calls. Here's what makes it significant.
Traditional WebRTC encryption (DTLS and SRTP) protects media in transit between peers, but in server-mediated scenarios—like video conferences using Selective Forwarding Units (SFUs)—the server can decrypt media to perform routing and optimization.
SFrame adds end-to-end encryption that prevents media from being decrypted even on intermediary servers. The SFU can still forward packets efficiently, but it can't inspect or modify the actual media content. Only the intended recipients can decrypt the audio and video.
This is critical for high-security applications: healthcare consultations handling patient data, legal discussions covered by attorney-client privilege, corporate board meetings discussing sensitive strategy. SFrame is recommended for any application where confidentiality requirements extend beyond basic transport security.
Forward Secrecy & Session Keys
One of WebRTC's standout security features is forward secrecy—a fresh encryption key is generated for every session. This means that even if current keys are compromised, past communications remain secure because they were encrypted with different, now-deleted keys.
DTLS handles encryption for data streams, SRTP (Secure Real-time Transport Protocol) handles encryption for media streams. Both generate ephemeral keys per session, ensuring that a breach today doesn't expose yesterday's conversations.
Compliance & Privacy
Security in 2025 is increasingly driven by regulatory compliance, not just best practices.
GDPR mandates encryption of personal data in transit, making WebRTC's mandatory encryption a baseline requirement for any application serving European users. Audio and video of identifiable individuals are considered personal data under GDPR.
HIPAA and SOC2 compliance require end-to-end encryption for telehealth and financial services. SFrame E2EE becomes necessary, not optional, for applications in these regulated industries.
WebRTC IP leak remains a privacy concern. Some browsers may inadvertently expose a user's real IP address through WebRTC even when using VPNs or anonymization tools. This can compromise user privacy, reveal geolocation, or leak personally identifiable information. Privacy-conscious applications need to implement protections against this.
The signaling channel—the mechanism that sets up WebRTC connections—should always use TLS (HTTPS or WSS) to prevent man-in-the-middle attacks and protect session metadata during connection setup.
Infrastructure Implications: Security vs Observability
From a relay operator's perspective, E2EE creates a fundamental tension: security requirements vs operational visibility.
When media is end-to-end encrypted with SFrame, relay servers cannot inspect the content. This is the point—it protects privacy and meets compliance requirements. But it also means operators lose the ability to perform quality diagnostics, detect codec issues, or troubleshoot stream problems by examining media content.
Traditional WebRTC troubleshooting involves analyzing RTCP reports, packet loss patterns, and sometimes inspecting frames to identify encoding problems. With E2EE, you can see packet-level metadata but not the content itself. Debugging becomes harder.
DTLS 1.3 support is mandatory for modern WebRTC infrastructure. Relay servers and TURN servers must upgrade to handle the new protocol version. Most operators have already completed this migration, but it's a reminder that security standards evolve and infrastructure must evolve with them.
Forward secrecy per-session keys mean there's no long-lived credential to cache or reuse. Each connection negotiates fresh keys, which adds a small computational overhead but provides the security guarantee that key compromise is limited to the current session only.
The balance is tricky: operators must provide strong security to meet compliance requirements and user expectations, while maintaining enough operational visibility to diagnose problems when they occur. The trend is clear—security and privacy are non-negotiable, and infrastructure must adapt to support them even when it makes operations more complex.
Trend 7 — Market Growth: $247.7 Billion Expansion
The WebRTC market isn't just growing—it's accelerating. Multiple research firms project extraordinary growth through 2033, driven by remote work normalization, telehealth adoption, IoT expansion, and the trends we've already discussed.
Market Size Projections (2025-2033)
Different research firms use different methodologies, which explains variance in estimates. But they all agree on one thing: growth is explosive.
Technavio projects the market will grow by USD 247.7 billion from 2025 to 2029, expanding at a CAGR of 62.6%. This is one of the highest growth rates in enterprise software.
Fortune Business Insights estimates the market at $9.56 billion in 2025, growing to $94.07 billion by 2032—a CAGR of 38.6%.
IMARC Group sizes the market at $11.6 billion in 2024, reaching $127.8 billion by 2033 with a CAGR of 30.3%.
The variance comes from how each firm defines "the WebRTC market." Some include only infrastructure and relay services. Others include CPaaS platforms, application development, and related services. Still others account for the entire value chain including devices, bandwidth, and support.
Regardless of which estimate you trust, the directional message is unmistakable: this market is growing faster than almost any other enterprise technology category.
Regional Adoption Patterns
North America holds 37.55% market share as of 2024, making it the current leader. Mature markets, high broadband penetration, and early adoption of remote work tools have driven WebRTC usage.
APAC is showing the fastest growth rate, fueled by rapid digitalization in India and Southeast Asia, expanding 5G networks in China and South Korea, and large populations of mobile-first users who leapfrog traditional desktop infrastructure.
The APAC opportunity is enormous but requires region-specific infrastructure. A WebRTC platform serving users in Jakarta, Manila, and Hanoi needs relay infrastructure in Southeast Asia—not just Tokyo or Singapore. Latency to users in Indonesia from a Singapore TURN server might be acceptable, but latency from Virginia is not.
EMEA shows steady growth with GDPR compliance driving demand for secure, privacy-preserving solutions. European enterprises prioritize data residency and encryption, making region pinning and E2EE capabilities differentiators in this market.
Industry Vertical Adoption
Telehealth has seen explosive growth. 54% of Americans had experienced a telehealth visit by 2024, and telehealth visits surged 38 times from pre-pandemic levels. While some expected a decline as pandemic restrictions eased, the convenience proved sticky—up to 30% of U.S. consultations are expected to remain virtual by 2026.
WebRTC is the technical foundation enabling browser-based telehealth. Patients join from a web browser without installing apps. Providers can conduct HIPAA-compliant video consultations without complex IT infrastructure.
Enterprise collaboration has normalized remote and hybrid work. The "return to office" trend never fully materialized at many companies. WebRTC powers the video conferencing and screen sharing tools that make distributed teams functional.
SMEs are adopting WebRTC solutions because of cost-effectiveness and scalability. Small businesses with geographically dispersed teams can't afford dedicated IT infrastructure, but they can use cloud-based WebRTC platforms that scale automatically and bill by usage.
Education has embraced virtual classrooms, breakout rooms, and screen sharing. While in-person instruction has resumed, hybrid and fully remote learning models remain common. WebRTC enables interactive educational experiences that aren't possible with one-way video broadcast.
Cloud Migration & Platform Consolidation
There's a clear shift from on-premise WebRTC infrastructure to cloud-based platforms. Organizations that previously ran self-hosted coturn servers are migrating to managed TURN services to reduce operational burden and improve reliability.
All-in-one CPaaS platforms are gaining traction. Instead of stitching together separate services for TURN relay, signaling, recording, and analytics, companies are consolidating on platforms that bundle these capabilities with predictable pricing and unified support.
The advantage of managed services is operational: no need to patch servers at 2 AM, no capacity planning guesswork, no multi-region deployment projects. The infrastructure scales automatically and bills by usage.
Self-hosted coturn remains popular for companies with specific compliance requirements or very large scale where dedicated infrastructure is cost-effective. But the median use case is shifting toward managed services.
Infrastructure Implications: Scaling for Exponential Growth
From an infrastructure operator's perspective, 62% CAGR creates massive scaling challenges.
Technical scaling: If traffic doubles year-over-year, infrastructure must more than double (to maintain headroom for spikes). This means continuous capacity planning, hardware procurement cycles, and network expansion.
Cost scaling: While revenue should grow with traffic, infrastructure costs aren't perfectly linear. At certain thresholds, you need bigger servers, additional regions, more robust network connectivity. Managing cost-per-GB as scale increases requires constant optimization.
Geographic expansion: Multi-region deployment is no longer optional—it's becoming the baseline expectation. Customers deploying globally expect relay infrastructure in APAC, EMEA, and the Americas at minimum.
TURN relay demand growing exponentially: As IoT adoption accelerates (where TURN is required 100% of the time, not 15-20%), relay traffic will grow faster than total WebRTC adoption. This changes infrastructure mix—more relay capacity needed relative to signaling and other services.
TCO advantage of managed TURN: A team running self-hosted coturn spends 15-20 hours per month on maintenance, monitoring, and troubleshooting. At $150-200/hour loaded engineer cost, that's $2,700-4,000 per month in opportunity cost—often more than a managed service would cost, and without the reliability, global distribution, or 24/7 support.
The market is expanding faster than most predicted. The infrastructure to support this growth must scale just as aggressively—and operators who can't keep pace will lose market share to those who can.
What These Trends Mean for Infrastructure Operators
We've covered seven trends shaping WebRTC in 2025. Now here's the perspective you won't find anywhere else: what do these trends actually mean for the infrastructure that makes WebRTC work?
No competitor writes about WebRTC from a TURN relay operator's viewpoint. They cover market trends and application features, but not the architectural, economic, and operational implications for the infrastructure layer. That's a blind spot—and a major one.
TURN Relay Architecture Implications
AI voice agents require global relay for sub-300ms latency. When a user in Singapore talks to an AI hosted in US-East, the relay path can't add more than 50-100ms or the interaction feels sluggish. This demands geographically distributed TURN servers with optimized inter-region connectivity.
It's not enough to have a TURN server in Singapore and another in Virginia. They need to be connected by a private, high-speed backbone that prioritizes latency over cost. Public internet routing can add 100-200ms for transcontinental connections during congestion. Private backbones avoid this.
AR/VR applications amplify this requirement. Cross-continent immersive collaboration needs sub-50ms network latency. The only way to achieve this reliably is private relay paths between TURN servers optimized for latency and jitter, not just throughput.
IoT deployments need always-on relay because devices sit behind carrier-grade NAT. Unlike video calls where TURN is a fallback, IoT requires TURN 100% of the time. This changes capacity planning—you're not sizing for 15-20% fallback traffic, you're sizing for 100% relay load.
MoQ adaptation means preparing for dual-protocol support. When MoQ matures in 2026+, relay infrastructure will need to handle both traditional WebRTC TURN and MoQ relay entities. The two protocols serve different use cases, so both will coexist rather than one replacing the other.
Bandwidth Economics & Codec Impact
AV1 adoption delivers 30-50% bandwidth savings at scale. For an infrastructure operator handling 10 petabytes of relay traffic per month, that could represent $100,000+ in monthly bandwidth cost reduction (depending on transit pricing). But AV1 adoption is gradual, not overnight, so cost reduction accrues slowly over 2025-2026.
Codec selection trade-offs affect infrastructure load differently. VP9 with SVC reduces bandwidth for group calls but increases CPU load on servers handling the forwarding logic. H.264/H.265 with hardware encoding reduces CPU but may increase bandwidth consumption. Operators must balance server costs (CPU, memory) against transit costs (bandwidth).
Traffic growth of 62% CAGR means bandwidth costs grow exponentially if not managed. Optimizing codec usage, upgrading to more efficient codecs as adoption allows, and negotiating volume discounts with transit providers become critical cost management strategies.
Egress fees at cloud providers can be prohibitive. If you're running TURN infrastructure on AWS, Azure, or GCP, egress (data leaving the cloud provider's network) can cost $0.05-$0.12 per GB. At petabyte scale, that's tens of thousands per month just in egress. Many operators are moving to colocation or bare-metal to eliminate egress fees entirely.
Security & Relay Challenges
E2EE prevents relay diagnostics. When SFrame encrypts media end-to-end, relay operators can see packet metadata (timing, size, destination) but not content. This makes troubleshooting codec issues, quality problems, or corruption significantly harder.
Traditional debugging involves inspecting frames to see if corruption occurred during encoding or transmission. With E2EE, you can't inspect frames—you can only infer problems from packet loss patterns and RTCP reports.
DTLS 1.3 migration requires infrastructure updates. TURN servers must support the new protocol version. Most operators completed this in early 2025, but it's a reminder that security standards evolve continuously and infrastructure must keep up.
Forward secrecy per-session keys mean no credential caching or reuse. Each connection negotiates fresh keys, adding computational overhead. At scale, this impacts CPU usage on TURN servers handling thousands of concurrent connections.
The balance is tricky: providing strong security to meet compliance and user expectations while maintaining operational visibility to diagnose and resolve issues quickly.
Regional Distribution & Data Residency
APAC fastest growth means infrastructure without strong APAC presence will struggle. A TURN provider with only North America and Europe coverage can't serve the fastest-growing market effectively. Latency from Jakarta to Frankfurt is 150-200ms—unacceptable for real-time applications.
GDPR and data residency requirements mean some customers need guarantees that media doesn't leave specific regions. A telehealth provider serving EU patients might require that all relay happens within EU data centers to comply with GDPR.
Region pinning becomes a differentiator. The ability to force all traffic for a specific customer or use case to relay through specific geographic regions addresses compliance requirements that are non-negotiable in regulated industries.
Multi-region deployment used to be a "nice to have" for better latency. In 2025, it's becoming a hard requirement for serving global customers and meeting compliance obligations.
Scaling for Market Growth
18 billion IoT devices plus 62% CAGR means infrastructure must scale aggressively and continuously. This isn't a one-time capacity addition—it's an ongoing procurement, deployment, and optimization cycle.
Auto-scaling and multi-region failover are becoming baseline expectations, not premium features. Customers expect infrastructure to handle traffic spikes without manual intervention and to fail over seamlessly if a region goes down.
Managed service advantages become more pronounced at scale. Running self-hosted coturn for a small deployment might make sense, but at scale, the operational complexity, multi-region coordination, and 24/7 monitoring requirements favor managed services.
TCO comparison is compelling: 15-20 hours per month of senior engineer time spent on TURN infrastructure (monitoring, patching, troubleshooting, scaling) costs $36,000-$50,000 per year in opportunity cost at typical senior engineer salaries. Many companies would save money and reduce risk by offloading this to a managed provider, even at $2,000-5,000/month.
How Metered Enables These Trends
Metered's infrastructure was built specifically to address these challenges:
31+ regions and 100+ PoPs provide the global distribution that AI, IoT, and XR applications require. Users in Tokyo, São Paulo, and Bangalore all connect to local TURN servers with low latency.
Private TURN backbone delivers the optimized relay paths critical for AI voice agents (<300ms latency requirement) and cross-continent AR/VR collaboration (<50ms latency requirement). Media relayed between continents travels over Metered's dedicated network, not the unpredictable public internet.
Sub-30ms global latency from anywhere in the world enables latency-sensitive applications that would be impossible with higher-latency infrastructure.
Premium bandwidth from local providers with direct peering maintains consistent quality even during network congestion. Settlement-free bandwidth (used by some competitors) degrades when the public internet is congested. Metered's paid bandwidth guarantees quality at all times.
Region pinning addresses GDPR and data residency requirements by allowing customers to force all relay traffic through specific geographic regions.
99.999% uptime SLA provides the reliability that mission-critical applications—telehealth, enterprise collaboration, financial services—demand. Five nines means less than 5 minutes of downtime per year.
The infrastructure that works for 2026's WebRTC trends isn't the same as what worked in 2020. The requirements have changed fundamentally, and operators who haven't adapted will struggle to serve the emerging use cases driving growth. You can test your TURN server to verify whether your current infrastructure meets these latency and connectivity benchmarks.
Conclusion — WebRTC's Transformative Year
2025 is the year WebRTC transitions from niche real-time communication technology to foundational internet infrastructure. AI integration is moving from experimental to production. Media over QUIC is emerging as a scalable broadcast solution. AV1 is beginning its gradual march toward mainstream adoption. IoT devices are adopting WebRTC at unprecedented scale. AR/VR applications are going mainstream. Security standards are strengthening to meet regulatory requirements. And the market is growing at 62% CAGR.
From an infrastructure operator's perspective, these trends demand robust, globally distributed TURN relay that can deliver sub-300ms latency for AI, sub-50ms latency for AR/VR, always-on relay for billions of IoT devices, and compliance-ready region pinning for regulated industries.
The workloads are more demanding. The scale is larger. The geographic distribution requirements are stricter. And the cost of failure—whether that's latency making AI conversations unnatural, or downtime breaking telehealth consultations—is higher than ever.
2026 will bring MoQ production maturity, broader AV1 hardware acceleration, and continued AI integration. The infrastructure requirements will only intensify. Operators who invest in global distribution, low-latency relay paths, and compliance capabilities now will have decisive advantages as these trends accelerate.
The infrastructure that powers WebRTC in 2026 isn't a commodity—it's a competitive differentiator that determines which applications can exist and which can't. See how Metered's global TURN infrastructure supports these trends with 31+ regions and sub-30ms latency.
FAQs — WebRTC Trends 2026
What are the biggest WebRTC trends in 2026?
The seven biggest trends are AI and machine learning integration (voice agents, real-time translation), Media over QUIC protocol emergence (combining WebRTC latency with HLS scale), codec evolution (AV1 bandwidth savings), IoT and edge computing convergence (18 billion devices), AR/VR/XR expansion (spatial audio, immersive experiences), security enhancements (DTLS 1.3, SFrame E2EE), and explosive market growth (62% CAGR, $247.7 billion expansion through 2029).
How is AI changing WebRTC?
AI enhances WebRTC with real-time language translation during video calls, machine learning-powered noise suppression that isolates voices from background sounds, video upscaling that improves low-resolution streams dynamically, sentiment analysis for customer service applications, and sign language translation for accessibility. The OpenAI Realtime API's WebRTC integration (announced December 2024) enables developers to build AI voice agents with sub-300ms latency for natural conversations.
What is Media over QUIC (MoQ)?
MoQ is a new streaming protocol developed at the IETF by engineers from Google, Meta, Cisco, Akamai, and Cloudflare. It solves streaming's "historical trilemma" by combining sub-second latency (like WebRTC), broadcast scale (like HLS/DASH), and architectural simplicity. Cloudflare launched the first MoQ relay network in 2025 across 330+ cities. MoQ complements WebRTC rather than competing—WebRTC for interactive communication, MoQ for scalable broadcast. Production readiness is expected in 2026+.
Is WebRTC secure in 2025?
Yes. WebRTC has mandatory encryption on all components (video, audio, data channels). The ecosystem migrated to DTLS 1.3 in February 2025, providing stronger cryptographic primitives and removing vulnerable legacy ciphers. SFrame end-to-end encryption is being standardized through IETF, preventing media decryption even on intermediary servers. Forward secrecy generates fresh encryption keys per session, ensuring compromised current keys can't decrypt past communications. GDPR, HIPAA, and SOC2 compliance requirements are driving adoption of these enhanced security measures.
What industries are adopting WebRTC?
Telehealth saw 54% of Americans use video consultations by 2024 (38× surge from pre-pandemic levels), with 30% expected to remain virtual by 2026. Enterprise collaboration platforms use WebRTC for distributed teams. SMEs adopt WebRTC for cost-effective communication among geographically dispersed teams. IoT devices (smart cameras, video doorbells, industrial sensors) use WebRTC for real-time monitoring. AR/VR applications use WebRTC for cross-platform immersive experiences. Education platforms use WebRTC for virtual classrooms and interactive learning.
What is the WebRTC market size in 2026?
Market size estimates vary by research firm methodology. Technavio projects USD 247.7 billion growth from 2025-2029 (62.6% CAGR). Fortune Business Insights estimates $9.56 billion in 2025 growing to $94.07 billion by 2032 (38.6% CAGR). IMARC Group sizes the market at $11.6 billion in 2024 reaching $127.8 billion by 2033 (30.3% CAGR). All reports agree on explosive growth driven by AI integration, IoT expansion, telehealth adoption, and remote work normalization.
What codecs does WebRTC support in 2026?
WebRTC supports VP8 (universal compatibility), VP9 (only codec with Scalable Video Coding for group calls), H.264 (maximum compatibility across devices), H.265/HEVC (hardware-accelerated efficiency on supported devices, Chrome 136 Beta added support), and AV1 (30-50% bandwidth savings but 5-10× slower encoding without hardware acceleration). In practice, VP8 and H.264 remain the workhorses handling most WebRTC traffic in 2025, with gradual AV1 adoption as hardware support improves.
For the official WebRTC specification, see the W3C WebRTC Recommendation (2025).




Top comments (1)
Thanks for reading. I hope you like the article