Remove 2018 Remove Cache Remove Latency Remove Servers
article thumbnail

How to use Server Timing to get backend transparency from your CDN

Speed Curve

Server-timing headers are a key tool in understanding what's happening within that black box of Time to First Byte (TTFB). Cue server-timing headers Historically, when looking at page speed, we've had the tendency to ignore TTFB when trying to optimize the user experience. I mean, why wouldn't we?

Servers 57
article thumbnail

Self-Host Your Static Assets

CSS Wizardry

Users might already have the file cached. If website-a.com links to [link] , and a user goes from there to website-b.com who also links to [link] , then the user will already have that file in their cache. Critical assets are far too valuable to leave on someone else’s servers. Risk: Slowdowns and Outages. to just 3.6s.

Cache 274
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

KeyCDN Launches POP in Helsinki

KeyCDN

With Helsinki being the capital and most populous municipality of Finland, it makes for a great edge server location. Although both countries are relatively close to one another, they are separated by a distance of approximately 500km, which adds up in terms of latency.

article thumbnail

Three Other Models of Computer System Performance: Part 2

ACM Sigarch

How many buffers are needed to track pending requests as a function of needed bandwidth and expected latency? Can one both minimize latency and maximize throughput for unscheduled work? The M/M/1 queue will show us a required trade-off among (a) allowing unscheduled task arrivals, (b) minimizing latency, and (c) maximizing throughput.

Systems 53
article thumbnail

Service Workers can save the environment!

Dean Hume

As data centers and volumes of servers have grown, so has the overall amount of electricity consumed around the world. Electricity used by servers doubled between 2000 and 2005 (and has continued growing ever since) from 12 billion to 23 billion kilowatt hours. Server Power Consumption (Source: Intel Labs 2008).

Energy 40