article thumbnail

Self-Host Your Static Assets

CSS Wizardry

Users might already have the file cached. If website-a.com links to [link] , and a user goes from there to website-b.com who also links to [link] , then the user will already have that file in their cache. On a slower, higher-latency connection, the story is much, mush worse. Penalty: Caching. Risk: Service Shutdowns.

Cache 274
article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Stuff The Internet Says On Scalability For July 20th, 2018

High Scalability

That means multiple data indirections mean multiple cache misses. Mark LaPedus : MRAM, a next-generation memory type, is being touted as a replacement for embedded flash and cache applications. Cliff Click : The JVM is very good at eliminating the cost of code abstraction, but not the cost of data abstraction. They are very expensive.

Internet 121
article thumbnail

How to use Server Timing to get backend transparency from your CDN

Speed Curve

Charlie Vazac introduced server timing in a Performance Calendar post circa 2018. Caching the base page/HTML is common, and it should have a positive impact on backend times. Key things to understand from your CDN Cache Hit/Cache Miss – Was the resource served from the edge, or did the request have to go to origin?

Servers 57
article thumbnail

KeyCDN Launches New POPs in 2021

KeyCDN

The image below shows a significant drop in latency once we've launched the new point of presence in Israel. In fact, latency has been reduced by almost 50%! With a total of 5 POPs in Oceania, this continent benefits from lower latency with every POP added. With a population of 2.5 What's next?

Latency 110
article thumbnail

Analyzing a High Rate of Paging

Brendan Gregg

1072-aws (xxx) 12/18/2018 _x86_64_ (16 CPU) avg-cpu: %user %nice %system %iowait %steal %idle 5.03 Reads usually have apps waiting on them; writes may not (write-back caching). biolatency From [bcc], this eBPF tool shows a latency histogram of disk I/O. This is a 64-Gbyte memory system, and 48 Gbytes is in the page cache.

Cache 105
article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP