Remove Exercise Remove Latency Remove Servers Remove Storage
article thumbnail

Migrating Critical Traffic At Scale with No Downtime?—?Part 1

The Netflix TechBlog

Replay Traffic Testing Replay traffic refers to production traffic that is cloned and forked over to a different path in the service call graph, allowing us to exercise new/updated systems in a manner that simulates actual production conditions. This approach has a handful of benefits.

Traffic 339
article thumbnail

A Decade of Dynamo: Powering the next wave of high-performance, internet-scale applications

All Things Distributed

percent availability in the event of a server, a rack of servers, or an Availability Zone failure. DynamoDB automatically re-distributes your data to healthy servers to ensure there are always multiple replicas of your data without you needing to intervene. Auto Scaling is on by default for all new tables and indexes.

Internet 128
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Evaluating the Evaluation: A Benchmarking Checklist

Brendan Gregg

sounds like a homework exercise of purely academic value. Networks, PCIe busses, CPU interconnects, memory busses, and storage devices (both throughput and IOPS), all have fixed limits. Once, during a proof of concept, a client reported that latency was unacceptably high for the benchmark: over one second for each request!

article thumbnail

Evaluating the Evaluation: A Benchmarking Checklist

Brendan Gregg

sounds like a homework exercise of purely academic value. Networks, PCIe busses, CPU interconnects, memory busses, and storage devices (both throughput and IOPS), all have fixed limits. Once, during a proof of concept, a client reported that latency was unacceptably high for the benchmark: over one second for each request!

article thumbnail

Amazon EC2 Cluster GPU Instances - All Things Distributed

All Things Distributed

For example, the most fundamental abstraction trade-off has always been latency versus throughput. Modern CPUs strongly favor lower latency of operations with clock cycles in the nanoseconds and we have built general purpose software architectures that can exploit these low latencies very well. Where to go from here?

AWS 136