Remove Cache Remove Latency Remove Performance Remove Traffic
article thumbnail

The Power of Caching: Boosting API Performance and Scalability

DZone

Caching is the process of storing frequently accessed data or resources in a temporary storage location, such as memory or disk, to improve retrieval speed and reduce the need for repetitive processing. Bandwidth optimization: Caching reduces the amount of data transferred over the network, minimizing bandwidth usage and improving efficiency.

Cache 246
article thumbnail

Migrating Critical Traffic At Scale with No Downtime?—?Part 1

The Netflix TechBlog

Migrating Critical Traffic At Scale with No Downtime — Part 1 Shyam Gala , Javier Fernandez-Ivern , Anup Rokkam Pratap , Devang Shah Hundreds of millions of customers tune into Netflix every day, expecting an uninterrupted and immersive streaming experience. This approach has a handful of benefits.

Traffic 339
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Consistent caching mechanism in Titus Gateway

The Netflix TechBlog

We introduce a caching mechanism in the API gateway layer, allowing us to offload processing from singleton leader elected controllers without giving up strict data consistency and guarantees clients observe. With traffic growth, a single leader node handling all request volume started becoming overloaded.

Cache 224
article thumbnail

Seeing through hardware counters: a journey to threefold performance increase

The Netflix TechBlog

A quick canary test was free of errors and showed lower latency, which is expected given that our standard canary setup routes an equal amount of traffic to both the baseline running on 4xl and the canary on 12xl. What’s worse, average latency degraded by more than 50%, with both CPU and latency patterns becoming more “choppy.”

Hardware 363
article thumbnail

Migrating Netflix to GraphQL Safely

The Netflix TechBlog

This blog post will share broadly-applicable techniques (beyond GraphQL) we used to perform this migration. And we definitely couldn’t replay test non-functional requirements like caching and logging user interaction. To determine customer impact, we could compare various metrics such as error rates, latencies, and time to render.

Traffic 353
article thumbnail

Crucial Redis Monitoring Metrics You Must Watch

Scalegrid

Redis® is an in-memory database that provides blazingly fast performance. This makes it a compelling alternative to disk-based databases when performance is a concern. You might already use ScaleGrid hosting for Redis hosting to power your performance-sensitive applications.

Metrics 130
article thumbnail

Redis vs Memcached in 2024

Scalegrid

In this comparison of Redis vs Memcached, we strip away the complexity, focusing on each in-memory data store’s performance, scalability, and unique features. Redis and Memcached both provide high performance with sub-millisecond response times. Choosing between Redis and Memcached hinges on specific application requirements.

Cache 130