article thumbnail

Zero Configuration Service Mesh with On-Demand Cluster Discovery

The Netflix TechBlog

These design principles led us to client-side load-balancing, and the 2012 Christmas Eve outage solidified this decision even further. There is a downside to fetching this data on-demand: this adds latency to the first request to a cluster.

Traffic 220
article thumbnail

USENIX LISA2021 Computing Performance: On the Horizon

Brendan Gregg

## References I've reproduced the talk references below, so you can click on links: - [Gregg 08] Brendan Gregg, “ZFS L2ARC,” [link] Jul 2008 - [Gregg 10] Brendan Gregg, “Visualizations for Performance Analysis (and More),” [link] 2010 - [Greenberg 11] Marc Greenberg, “DDR4: Double the speed, double the latency? Ford, et al., “TCP

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Towards a Unified Theory of Web Performance

Alex Russell

Tim Berners-Lee tweets that 'This is for everyone' at the 2012 Olympic Games opening ceremony using the NeXT computer he used to build the first browser and web server. The chief effect of the architectural difference is to shift the distribution of latency within the loop. Improving latency for one scenario can degrade it in another.

article thumbnail

Stuff The Internet Says On Scalability For March 22nd, 2019

High Scalability

µs of replication latency on lossy Ethernet, which is faster than or comparable to specialized replication systems that use programmable switches, FPGAs, or RDMA.". BrentToderian : What city went from 14% of all trips by bike in 2001, to 22% by 2012, then leaped to 30% in 3 years by 2015, & 35% by 2018? We achieve 5.5

Internet 134
article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

New Year’s Updates

John McCalpin

This week I updated three of those posts — two really old ones (primarily of interest to computer architecture historians), and one from 2018: July 2012: Local and Remote Memory Latency on AMD Processors in 2-socket and 4-socket servers. December 2013: Notes on Memory Bandwidth on the Xeon Phi (Knights Corner) Coprocessor.

Latency 40
article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP