Remove Latency Remove Scalability Remove Technology Remove Tuning
article thumbnail

Migrating Critical Traffic At Scale with No Downtime?—?Part 1

The Netflix TechBlog

Migrating Critical Traffic At Scale with No Downtime — Part 1 Shyam Gala , Javier Fernandez-Ivern , Anup Rokkam Pratap , Devang Shah Hundreds of millions of customers tune into Netflix every day, expecting an uninterrupted and immersive streaming experience. Logging is selective to cases where the old and new responses do not match.

Traffic 339
article thumbnail

Rebuilding Netflix Video Processing Pipeline with Microservices

The Netflix TechBlog

This architecture shift greatly reduced the processing latency and increased system resiliency. We expanded pipeline support to serve our studio/content-development use cases, which had different latency and resiliency requirements as compared to the traditional streaming use case. divide the input video into small chunks 2.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Growth Engineering at Netflix- Creating a Scalable Offers Platform

The Netflix TechBlog

Stay tuned for more details on this, as well as more details on the internals of the new SKU Platform in one of our upcoming blog posts. Lower latency as a result of fewer service calls, which means fewer errors for our visitors. The next step is to invest further into self-service and support rule changes via a SKU UI.

article thumbnail

Automated observability, security, and reliability at scale

Dynatrace

Efficient environment configuration at scale One of software engineers’ most significant challenges is managing the numerous tools and technologies required for the software product lifecycle. Stay tuned for more examples and easy-to-adopt automations provided in our public Github project.

article thumbnail

The Netflix Cosmos Platform

The Netflix TechBlog

It supports both high throughput services that consume hundreds of thousands of CPUs at a time, and latency-sensitive workloads where humans are waiting for the results of a computation. The third generation, called Reloaded , has been online for about seven years and has proven to be stable and massively scalable.

article thumbnail

Bulldozer: Batch Data Moving from Data Warehouse to Online Key-Value Stores

The Netflix TechBlog

Usually Data scientists and engineers write Extract-Transform-Load (ETL) jobs and pipelines using big data compute technologies, like Spark or Presto , to process this data and periodically compute key information for a member or a video. The data warehouse is not designed to serve point requests from microservices with low latency.

Latency 243
article thumbnail

An empirical guide to the behavior and use of scalable persistent memory

The Morning Paper

An empirical guide to the behavior and use of scalable persistent memory , Yang et al., higher latency and lower bandwidth)… We have found the actual behavior of Optane DIMMs to be more complicated and nuanced than the "slower, persistent DRAM" label would suggest. The read latency for Optane is 2x-3x higher than DRAM.