Remove Infrastructure Remove Latency Remove Traffic Remove Video
article thumbnail

Rebuilding Netflix Video Processing Pipeline with Microservices

The Netflix TechBlog

The Netflix video processing pipeline went live with the launch of our streaming service in 2007. This architecture shift greatly reduced the processing latency and increased system resiliency. This introductory blog focuses on an overview of our journey. We moved from centralized linear encoding to distributed chunk-based encoding.

article thumbnail

Migrating Netflix to GraphQL Safely

The Netflix TechBlog

We could also swap out the implementation of a field from GraphQL Shim to Video API with federation directives. The control group’s traffic utilized the legacy Falcor stack, while the experiment population leveraged the new GraphQL client and was directed to the GraphQL Shim. How does it work?

Traffic 353
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Netflix Video Quality at Scale with Cosmos Microservices

The Netflix TechBlog

Moorthy and Zhi Li Introduction Measuring video quality at scale is an essential component of the Netflix streaming pipeline. Perceptual quality measurements are used to drive video encoding optimizations , perform video codec comparisons , carry out A/B testing and optimize streaming QoE decisions to mention a few.

Media 171
article thumbnail

Building Netflix’s Distributed Tracing Infrastructure

The Netflix TechBlog

Investigating a video streaming failure consists of inspecting all aspects of a member account. Now let’s look at how we designed the tracing infrastructure that powers Edgar. This insight led us to build Edgar: a distributed tracing infrastructure and user experience. which is difficult when troubleshooting distributed systems.

article thumbnail

Service level objectives: 5 SLOs to get started

Dynatrace

Note : you might hear the term latency used instead of response time. Both latency and response time are critical to ensure reliability. Latency typically refers to the time it takes for a single request to travel from its source to its destination. Latency primarily focuses on the time spent in transit.

Latency 182
article thumbnail

Service level objective examples: 5 SLO examples for faster, more reliable apps

Dynatrace

Note : you might hear the term latency used instead of response time. Both latency and response time are critical to ensure reliability. Latency typically refers to the time it takes for a single request to travel from its source to its destination. Latency primarily focuses on the time spent in transit.

Traffic 173
article thumbnail

How to Reduce Your CDN Infrastructure Expenses

IO River

Common Infrastructure ExpensesYour first step in optimizing CDN expenses isn’t to look for the best-priced solution but to remember that a cheaper price isn’t always the best deal. If your traffic is mostly static, you may be able to meet all your needs with a less expensive CDN that provides content distribution services.