article thumbnail

Migrating Critical Traffic At Scale with No Downtime?—?Part 1

The Netflix TechBlog

Migrating Critical Traffic At Scale with No Downtime — Part 1 Shyam Gala , Javier Fernandez-Ivern , Anup Rokkam Pratap , Devang Shah Hundreds of millions of customers tune into Netflix every day, expecting an uninterrupted and immersive streaming experience. This approach has a handful of benefits.

Traffic 339
article thumbnail

Maximize user experience with out-of-the-box service-performance SLOs

Dynatrace

This article explores SLOs for service performance. According to the Google Site Reliability Engineering (SRE) handbook, monitoring the four golden signals is crucial in delivering high-performing software solutions. SLOs, as a measure of service quality, can track the related availability, reliability, and performance.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

The Power of Caching: Boosting API Performance and Scalability

DZone

Benefits of Caching Improved performance: Caching eliminates the need to retrieve data from the original source every time, resulting in faster response times and reduced latency. Reduced server load: By serving cached content, the load on the server is reduced, allowing it to handle more requests and improving overall scalability.

Cache 246
article thumbnail

What are quality gates? How to use quality gates to deliver better software at speed and scale

Dynatrace

Benefits of quality gates Quality gates provide several advantages to organizations, including the following: Optimized software performance : Quality gates assess code at different SDLC stages and ensure that only high-quality code progresses. Several tools can be used to collect metrics in load/performance testing.

Speed 212
article thumbnail

FIFO vs. LIFO: Which Queueing Strategy Is Better for Availability and Latency?

DZone

As an engineer, you probably know that server performance under heavy load is crucial for maintaining the availability and responsiveness of your services. But what happens when traffic bursts overwhelm your system? Queueing requests is a common solution, but what's the best approach: FIFO or LIFO?

Strategy 141
article thumbnail

Bending pause times to your will with Generational ZGC

The Netflix TechBlog

Reduced tail latencies In both our GRPC and DGS Framework services, GC pauses are a significant source of tail latencies. Each of these errors is a canceled request resulting in a retry so this reduction further reduces overall service traffic by this rate: Errors rates per second. There is no best garbage collector.

Latency 228
article thumbnail

Event-Based Autoscaling: Ensuring Smooth Operations on Your Peak Days

DZone

These organizations face a common challenge – how much infrastructure do they need to ensure optimal performance without overprovisioning – which can become very costly, very quickly. Even retail giants like Amazon have faced customer dissatisfaction during events like Prime Day when the website couldn't handle the traffic.

Retail 169