article thumbnail

Migrating Critical Traffic At Scale with No Downtime?—?Part 1

The Netflix TechBlog

Migrating Critical Traffic At Scale with No Downtime — Part 1 Shyam Gala , Javier Fernandez-Ivern , Anup Rokkam Pratap , Devang Shah Hundreds of millions of customers tune into Netflix every day, expecting an uninterrupted and immersive streaming experience. This approach has a handful of benefits.

Traffic 339
article thumbnail

What are quality gates? How to use quality gates to deliver better software at speed and scale

Dynatrace

Before a new version of the application is deployed, the software is subject to a series of load tests that evaluate capacity and performance under a series of simulated traffic and application demands. These metrics are latency, traffic, errors, and saturation, all of which must be key considerations when curating user experience.

Speed 209
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

FIFO vs. LIFO: Which Queueing Strategy Is Better for Availability and Latency?

DZone

But what happens when traffic bursts overwhelm your system? As an engineer, you probably know that server performance under heavy load is crucial for maintaining the availability and responsiveness of your services. Queueing requests is a common solution, but what's the best approach: FIFO or LIFO?

Strategy 141
article thumbnail

Bending pause times to your will with Generational ZGC

The Netflix TechBlog

Reduced tail latencies In both our GRPC and DGS Framework services, GC pauses are a significant source of tail latencies. Each of these errors is a canceled request resulting in a retry so this reduction further reduces overall service traffic by this rate: Errors rates per second. There is no best garbage collector.

Latency 198
article thumbnail

Event-Based Autoscaling: Ensuring Smooth Operations on Your Peak Days

DZone

This poses a significant challenge for businesses since miscalculations can lead to latency, lost customers, and significant financial losses, even as much as hundreds of thousands of dollars per minute. Remember when the Game of Thrones spinoff had technical difficulties during its premiere ?

Retail 169
article thumbnail

The Power of Caching: Boosting API Performance and Scalability

DZone

Benefits of Caching Improved performance: Caching eliminates the need to retrieve data from the original source every time, resulting in faster response times and reduced latency. Reduced server load: By serving cached content, the load on the server is reduced, allowing it to handle more requests and improving overall scalability.

Cache 246
article thumbnail

Uber’s Big Data Platform: 100+ Petabytes with Minute Latency

Uber Engineering

To accomplish this, Uber relies heavily on making data-driven decisions at every level, from forecasting rider demand during high traffic events to identifying and addressing bottlenecks … The post Uber’s Big Data Platform: 100+ Petabytes with Minute Latency appeared first on Uber Engineering Blog.

Big Data 109