Remove Latency Remove Traffic Remove Video
article thumbnail

Rebuilding Netflix Video Processing Pipeline with Microservices

The Netflix TechBlog

The Netflix video processing pipeline went live with the launch of our streaming service in 2007. This architecture shift greatly reduced the processing latency and increased system resiliency. For example, in Reloaded the video quality calculation was implemented inside the video encoder module.

article thumbnail

Migrating Critical Traffic At Scale with No Downtime?—?Part 2

The Netflix TechBlog

Migrating Critical Traffic At Scale with No Downtime — Part 2 Shyam Gala , Javier Fernandez-Ivern , Anup Rokkam Pratap , Devang Shah Picture yourself enthralled by the latest episode of your beloved Netflix series, delighting in an uninterrupted, high-definition streaming experience. This is where large-scale system migrations come into play.

Traffic 287
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Migrating Netflix to GraphQL Safely

The Netflix TechBlog

We could also swap out the implementation of a field from GraphQL Shim to Video API with federation directives. The control group’s traffic utilized the legacy Falcor stack, while the experiment population leveraged the new GraphQL client and was directed to the GraphQL Shim. How does it work?

Traffic 358
article thumbnail

Netflix Video Quality at Scale with Cosmos Microservices

The Netflix TechBlog

Moorthy and Zhi Li Introduction Measuring video quality at scale is an essential component of the Netflix streaming pipeline. Perceptual quality measurements are used to drive video encoding optimizations , perform video codec comparisons , carry out A/B testing and optimize streaming QoE decisions to mention a few.

Media 180
article thumbnail

So many bad takes?—?What is there to learn from the Prime Video microservices to monolith story

Adrian Cockcroft

Then they tried to scale it to cope with high traffic and discovered that some of the state transitions in their step functions were too frequent, and they had some overly chatty calls between AWS lambda functions and S3. This is only one of many microservices that make up the Prime Video application. Finally, what were they building?

article thumbnail

Edgar: Solving Mysteries Faster with Observability

The Netflix TechBlog

Edgar captures 100% of interesting traces , as opposed to sampling a small fixed percentage of traffic. Telltale provides Edgar with latency benchmarks that indicate if the individual trace’s latency is abnormal for this given service. Is this an anomaly or are we dealing with a pattern?

Latency 299
article thumbnail

Seamlessly Swapping the API backend of the Netflix Android app

The Netflix TechBlog

As an example, to render the screen shown here, the app sends a query that looks like this: paths: ["videos", 80154610, "detail"] A path starts from a root object , and is followed by a sequence of keys that we want to retrieve the data for. Instead, it is part of a different path : [videos, <id>, similars].

Latency 243