Three years ago, Netflix asked a deceptively simple question: What would it take to stream live events with the same quality, scale, and reliability as our on-demand catalog?
What followed wasn’t a moonshot. It was a methodical, multi-year buildout that turned Netflix into a serious live platform that now supports everything from comedy specials and NFL games to record-setting boxing matches and WWE Raw. It hasn’t always gone smoothly. Events like Tyson vs. Paul came with their share of public glitches and technical backlash, but Netflix kept investing and evolving.
This article explores Netflix’s architectural choices, what makes them effective, and how smaller platforms can adapt, simplify, or skip certain elements based on their own needs. Because while most services won’t replicate Netflix end-to-end, there’s still a lot to learn from how they built it.
Honestly, if you’re a video engineer, you’re better off reading Netflix’s excellent post, Behind the Streams: Three Years Of Live at Netflix. Part 1, from which I derived this article. On the other hand, if you’re a bit less technical, I take a simpler swing at the same material. Also, at the end, I identify the takeways that streaming producers smaller than Netflix can actually use for their own benefit.
Contents
What Makes Live Streaming Different
Most of us know this, but Netflix started by detailing why live streaming is different than VOD, which has obviously been Netflix’s bread and butter since it’s DVD-based inception. I’ll let a picture be worth a thousand words and present Figure 1.

If anything, Figure 1 understates perhaps the biggest issue of all: delivery. As you’ll read, with its extensive network of edge servers around the world, Netflix’s internal CDN has a proven ability to deliver VOD video to massive audiences globally, 24/7. Unfortunately, much of this delivers little or no value for live streaming.
From Stand-Up to Scale: Netflix’s Live Evolution
Netflix’s first live event, Chris Rock: Selective Outrage, launched just nine months after the kickoff of the company’s live streaming architecture. It was an ambitious sprint, but behind the scenes, the team knew this would be more than a one-off. Supporting live at Netflix scale required foundational changes, not just in delivery, but across ingest, encoding, failover, playback logic, and real-time telemetry.
Rather than build a separate system, Netflix integrated live into its existing platform. The result: a unified pipeline from production to screen, where live and on-demand share infrastructure but differ in execution.

Cloud-Based Encoding and a Custom Origin
Netflix processes its contribution feed via AWS MediaConnect and MediaLive for cloud-based transcoding. As Netflix explained, using standard AWS services may seem surprising for a company of Netflix’s size, but it reflects a practical decision: flexibility and speed were more important than squeezing every last bit of encoding efficiency.

This decision informs the perennial argument between software-based and hardware-based transcoding. Software-based is much more expensive, and can deliver higher top-end quality, or better quality than hardware can at the same bitrates. GPU or ASIC-based hardware transcoding is cheaper but can’t achieve the same quality as software. If you’re transcoding the Super Bowl, or even the Tyson/Paul fight, encoding cost doesn’t matter, but you need the highest quality signal at the lowest possible bitrate. If, like Twitch, you’re transcoding hundreds of individual gamers broadcasting to dozens of viewers, hardware transcoding is the best option, even if your parent company owns the world’s largest cloud encoding platform. But, I digress.
Returning to the Netflix architecture, Netflix built a custom packager to better integrate with its delivery and playback systems. Though the GPAC packager isn’t explicitly identified in the Netflix article, it appears that Netflix used GPAC, in whole or in part, for this function. Netflix also built Live Origin, which is a custom-built origin service with strict guarantees on segment availability and timing. This origin helps manage the sensitivity of live playback, ensuring smooth segment handoffs and reducing the risk of interruptions.
Delivery on Open Connect, But Live
Netflix has long operated its own CDN, Open Connect, for on-demand video. The same network now supports live delivery, but with important differences.
Before diving into those differences, let’s explore what Open Connect is and how it works. There are two basic components. Figure 4 shows the Open Connect Servers located in ISPs or Internet Exchanges around the globe. These are physical appliances owned and operated by Netflix; the hosting location simply provides rack space, power, and connectivity.
The benefit for ISPs and exchanges is reduced backbone traffic and improved service quality for Netflix viewers. Instead of pulling each stream for each viewer from the origin server via the public internet, popular content is delivered by local servers, often from within the same city or facility as the viewer, and served from there. Content can be pre-delivered or simply stored on the server after delivery to the first local viewer. The quid pro quo for Netflix is free hosting for its servers, which translates to cheaper delivery and better quality of service to local viewers around the globe.

Figure 5 shows the dedicated Open Connect backbone, a private, high-throughput network that connects Netflix’s cloud infrastructure (e.g., AWS regions used for live encoding) to these edge servers. This backbone ensures the timely and reliable transfer of media assets, whether on-demand files or live video segments, to those appropriate edge locations.

You can imagine the huge benefits Open Connect delivers for video on demand, which can either be pre-cached or cached after first playback, neither of which is possible for live. Netflix doesn’t provide much technical detail about how Open Connect was adapted to support real-time live delivery. The only specifics come in a single line: “By sharing capacity across on-demand and Live viewership we improve utilization, and by caching past Live content on the same servers used for on-demand streaming, we can easily enable catch-up viewing.”
What’s left unsaid is how Netflix handles real-time segment propagation to 18,000 edge servers in a live context, where there’s no opportunity to pre-cache and timing is critical.
What we can safely infer is this: for live playback, the Open Connect Backbone carries the operational burden, transporting segments from AWS origins to the edge as they’re produced. That’s a major contrast from VOD, where the embedded servers do the heavy lifting, serving cached content with minimal real-time dependency on the backbone. Perhaps Netflix will tell us more in a future blog post.
Encoding and Delivery Format
Netflix then outlined some of the core technical decisions behind its Live playback stack. First up: delivery. Rather than chase low-latency edge cases like WebRTC, SRT, or the emerging MoQ standard, Netflix stuck with HTTPS-based streaming, think HLS or CMAF over HTTP. That means this isn’t a sub-second system, but it should work across nearly every device, browser, and network configuration worldwide.
From a codec standpoint, Netflix uses both AVC and HEVC, encoding each live stream into multiple quality levels from SD to 4K. AVC ensures backwards compatibility across older devices, while HEVC delivers better compression efficiency at higher resolutions. The article doesn’t mention content-aware encoding or per-title ladder optimization. MediaLive supports QVBR, a CRF-like per-title function, so it’s possible that Netflix uses some form of dynamic quality control, but there’s no confirmation either way. AV1 wasn’t available when Netflix started this journey, but AWS integrated AV1 output in September 2024; it will be interesting to see if Netflix starts deploying AV1 over its next few events.
Netflix deftly avoided detailing its end-to-end latency, perhaps the single most intriguing detail for many live event producers. Instead, the blog simply says, we “use a 2-second segment duration to balance compression efficiency, infrastructure load, and latency. While prioritizing streaming quality and playback stability, we have also achieved industry-standard latency from camera to device, and continue to improve it.”
Netflix made a few notable choices in how it delivers streams to different devices. Manifests are served from the cloud instead of the CDN, which allows Netflix to personalize stream settings for each device. Instead of periodic polling, manifests use a segment template that maps wall-clock time to segment URLs. This minimizes network chatter and reduces load on both CDN and device, especially on constrained hardware like smart TVs. During playback, the client can adapt not just the bitrate but also the CDN server, optimizing for quality while avoiding stalls.
Orchestrating The Event
The first half of Netflix’s article described the video pipeline, including ingest, encoding, and delivery. For the most part, it’s straightforward; choose a codec, encode and package, and deliver it to the CDN. Next, Netflix covered the orchestration and observability layer: the control systems that manage how live events are launched, scaled, and kept stable in real time. Although brief, this section outlines much of the secret sauce behind what Netflix has operationalized.
It starts with coordination. Netflix runs dozens of cloud services to handle playback setup, session personalization, and metrics collection. These services see the biggest spikes not during the stream, but right before it, when thousands or millions of users hit Play within seconds. To absorb that load, Netflix distributes the control infrastructure across multiple AWS regions, shifting traffic as needed based on geography and demand.
Equally important is Netflix’s observability stack. With visibility into nearly every component, from encoding to CDN to the device itself, Netflix processes up to 38 million telemetry events per second. These metrics include when a viewer presses Play, how long it takes for the video to start, whether it buffers, what bitrate is delivered, and whether playback completes successfully. Data flows through internal tools like Atlas, Mantis, and Lumen, along with open-source systems like Kafka and Druid, and is presented in real-time dashboards used by operational teams during major events.
This data allows Netflix to spot issues as they emerge, isolate them quickly, and take corrective action. If a segment is missing, a CDN node is overloaded, or a specific smart TV is failing to adapt, Netflix knows about it within seconds and can respond. Netflix sets up dedicated Control Centers during large events, where ops teams watch key metrics and coordinate fixes in real time. The goal isn’t just to stream, but to make sure that stream holds up under real-world conditions, across every device and region.
Lessons Learned
Netflix wraps up with a list of lessons learned. Here’s the boiled-down version.
-
Extensive testing: Netflix couldn’t rely on predictable on-demand traffic to validate Live workflows, so they built internal test streams and load generators. These tools simulate 100,000+ play starts per second and include failure injection across encoders, networks, and cloud regions.
-
Regular practice: Netflix treats every Live event as a chance to improve under real conditions. They run chaos tests, train ops teams, and maintain a regular cadence of diverse events to build muscle without risking untested launches.
-
Viewership predictions: Netflix combines pre-launch forecasts with real-time viewership modeling to scale infrastructure in advance. This lets them pre-warn ISPs, shift cloud resources, and react before problems hit most users.
-
Graceful degradation: When demand exceeds capacity, Netflix sheds lower-priority traffic and disables features like personalization or high-bitrate streams. Load testing includes these scenarios to ensure fallback behaviors work as intended.
-
Retry storms: User-initiated retries after minor playback failures can multiply load by 10x. Netflix uses server-guided backoff and regional traffic rebalancing to absorb the spike without cascading failures.
-
Contingency planning: Netflix doesn’t troubleshoot on the fly during big events; they rehearse failure responses in advance. Dedicated launch rooms and Game Day drills prepare cross-functional teams to escalate and resolve issues fast.
-
Post-event analysis: Every stream is followed by data reviews, A/B tests, and consumer feedback analysis. These insights have led to real improvements, like cutting latency by 10 seconds without harming quality.
So What Can You Actually Use?
Most platforms can’t build their own CDN or process tens of millions of telemetry events per second. But Netflix’s approach isn’t about scale alone — it’s about smart engineering and tight operational control. Here’s how different types of organizations can apply the right lessons for their size and scope:
Tier 1: Small & Simple (Live as a Feature)
Typical use case: Event producers, educators, small creators
Tools: Vimeo, YouTube Live, StreamYard
Why: Your priority is ease of use and reliability. You don’t need deep control — you need a provider that works out of the box.
What to apply from Netflix:
- Communicate proactively during issues
- Perform basic post-event reviews
- Choose a vendor with simple fallback options
Tier 2: Mid-Scale SaaS Stack (Live as a Product)
Typical use case: Niche streaming platforms, virtual events, branded apps
Tools: Mux, Wowza, AWS Elemental
Why: You want API access, some branding control, and data you can act on. You’re not managing your own CDN, but you want to optimize around it.
What to apply from Netflix:
- Serve manifests from your backend for session control
- Run synthetic load tests before major events
- Monitor real-time playback health and establish simple fallback paths
Tier 3: DIY Platform (Live as Core Business)
Typical use case: FAST platforms, sports leagues, high-volume streamers
Tools: Your own encoding, multi-CDN integration, telemetry pipelines
Why: Live is central to your offering. You’re responsible for QoE, reliability, and scale — and need to prove it every day.
What to apply from Netflix:
- Use or build an origin that supports tight read/write SLAs
- Generate cloud-based manifests with per-session logic
- Build real-time telemetry dashboards and chaos test your stack
- Coordinate with CDN and cloud partners ahead of known peaks
Tier 4: Global Scale (Live as Strategic Lever)
Typical use case: YouTube, Prime Video, Disney+
Why: You own the full delivery chain and treat live as a core innovation area.
What to apply from Netflix: I’m doubting any of you made it this far, but if so, since you’re already operating at this level, your challenge isn’t what to copy. It’s what to innovate next.
Final Thought
Netflix didn’t set out to reinvent live streaming. Rather, they set out to make it reliable, scalable, and invisible to the user. Their biggest success isn’t in chasing cutting-edge latency or proprietary protocols. It’s in integrating live into a platform already trusted by hundreds of millions and delivering at that same level of quality and scale.
For the rest of us, the lesson isn’t “build a CDN.” It’s think strategically, design for failure, and optimize for what your viewers actually need.
Streaming Learning Center Where Streaming Professionals Learn to Excel
