Remove AWS Remove Cache Remove Latency Remove Presentation
article thumbnail

Consistent caching mechanism in Titus Gateway

The Netflix TechBlog

In the time since it was first presented as an advanced Mesos framework, Titus has transparently evolved from being built on top of Mesos to Kubernetes, handling an ever-increasing volume of containers. This blog post presents how our current iteration of Titus deals with high API call volumes by scaling out horizontally.

Cache 224
article thumbnail

Jamstack CMS: The Past, The Present and The Future

Smashing Magazine

Jamstack CMS: The Past, The Present and The Future. Jamstack CMS: The Past, The Present and The Future. Twitch developer documentation hosted on AWS, edited on CloudCannon. When we talk about static site generators, incremental regeneration, or instant cache invalidation, it’s enough to make the layman’s eyes glaze over.

Ecommerce 139
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

How to use Server Timing to get backend transparency from your CDN

Speed Curve

Caching the base page/HTML is common, and it should have a positive impact on backend times. Key things to understand from your CDN Cache Hit/Cache Miss – Was the resource served from the edge, or did the request have to go to origin? Latency – How much time does it take to deliver a packet from A to B.

Servers 57
article thumbnail

Supercomputing Predictions: Custom CPUs, CXL3.0, and Petalith Architectures

Adrian Cockcroft

Clouds using Ethernet that are multipath optimized using libfabric and features like EFA on AWS are going to be increasingly competitive, and Ethernet will replace other interconnects between racks. I presented a keynote for Sun at Supercomputing 2003 in Phoenix Arizona and included the slide shown below.

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

Analyzing a High Rate of Paging

Brendan Gregg

1072-aws (xxx) 12/18/2018 _x86_64_ (16 CPU) avg-cpu: %user %nice %system %iowait %steal %idle 5.03 Reads usually have apps waiting on them; writes may not (write-back caching). biolatency From [bcc], this eBPF tool shows a latency histogram of disk I/O. This is a 64-Gbyte memory system, and 48 Gbytes is in the page cache.

Cache 105
article thumbnail

How We Optimized Performance To Serve A Global Audience

Smashing Magazine

Large preview ) The graph presents the LCP values for the 75th percentile of our users visiting the Route Landing Page group. It also opens up the possibility for more effective use of caching strategies, potentially enhancing load times further. LCP seconds over time.