article thumbnail

USENIX LISA2021 Computing Performance: On the Horizon

Brendan Gregg

AWS Graviton2); for memory with the arrival of DDR5 and High Bandwidth Memory (HBM) on-processor; for storage including new uses for 3D Xpoint as a 3D NAND accelerator; for networking with the rise of QUIC and eXpress Data Path (XDP); and so on. Ford, et al., “TCP

article thumbnail

Self-Host Your Static Assets

CSS Wizardry

This is exactly what Rawgit did in October 2018, yet (at the time of writing) a crude GitHub code search still yielded over a million references to the now-sunset service, and almost 20,000 live sites are still linking to it! Penalty: Network Negotiation. On a slower, higher-latency connection, the story is much, mush worse.

Cache 274
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

bpftrace (DTrace 2.0) for Linux 2018

Brendan Gregg

eBPF was created by Alexei Starovoitov while at PLUMgrid (he's now at Facebook) as a generic in-kernel virtual machine, with software defined networks as the primary use case. Here's key differences as of August 2018: Type DTrace bpftrace. It's shaping up to be a DTrace version 2.0: eBPF does more. variable this->name $name.

C++ 110
article thumbnail

Stuff The Internet Says On Scalability For March 22nd, 2019

High Scalability

µs of replication latency on lossy Ethernet, which is faster than or comparable to specialized replication systems that use programmable switches, FPGAs, or RDMA.". BrentToderian : What city went from 14% of all trips by bike in 2001, to 22% by 2012, then leaped to 30% in 3 years by 2015, & 35% by 2018? We achieve 5.5

Internet 134
article thumbnail

How to use Server Timing to get backend transparency from your CDN

Speed Curve

Charlie Vazac introduced server timing in a Performance Calendar post circa 2018. The use of server-timing headers by content delivery networks closes a big gap. Latency – How much time does it take to deliver a packet from A to B. In practice, it looks something like this: server-timing: processing_time; dur=123.4;

Servers 58
article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP