article thumbnail

Percentiles don’t work: Analyzing the distribution of response times for web services

Adrian Cockcroft

There is no way to model how much more traffic you can send to that system before it exceeds it’s SLA. Every opportunity for delay due to more work than the best case or more time waiting than the best case increases the latency and they all add up and create a long tail. Mu is the mean of each component, the latency.

Lambda 98
article thumbnail

A 5G future

O'Reilly

I don’t need more bandwidth for video conferences or movies, but I would like to be able to download operating system updates and other large items in seconds rather than minutes. There are impressive estimates for latency for 5G, but reality has a tendency to be harsh on such predictions. Upcoming events.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Why OpenStack is like a Crowdfunded Viking Movie

VoltDB

An opening scene involving a traffic jam of Viking boats and a musical number (“Love Can’t Afjord to wait”). Earlier this month I spent a week at the OpenStack conference. But once you realize that there are, in fact, several OpenStack conferences happening at the same time and place, it makes more sense. Vikings fight zombies.

article thumbnail

Why OpenStack is like a Crowdfunded Viking Movie

VoltDB

An opening scene involving a traffic jam of Viking boats and a musical number (“Love Can’t Afjord to wait”). Earlier this month I spent a week at the OpenStack conference. But once you realize that there are, in fact, several OpenStack conferences happening at the same time and place, it makes more sense. Vikings fight zombies.

article thumbnail

Revisiting “Serverless Architectures”

The Symphonia

I also rewrote the section on Startup Latency since Cold Starts are one of the big “FUD” areas of Serverless. I was glad to be able to talk about Amazon’s automated traffic shifting / canary releases. Right now there’s been one conference and there are a handful of meetups around the world.

article thumbnail

Consistent caching mechanism in Titus Gateway

The Netflix TechBlog

In that scenario, the system would need to deal with the data propagation latency directly, for example, by use of timeouts or client-originated update tracking mechanisms. With traffic growth, a single leader node handling all request volume started becoming overloaded. The cache is kept in sync with the current leader process.

Cache 224
article thumbnail

The Speed of Time

Brendan Gregg

A Cassandra database cluster had switched to Ubuntu and noticed write latency increased by over 30%. Since instances of both CentOS and Ubuntu were running in parallel, I could collect flame graphs at the same time (same time-of-day traffic mix) and compare them side by side. in total.

Speed 126