Cache-Control for Civilians

CSS Wizardry

To this end, having a solid caching strategy can make all the difference for your visitors. ?? How is your knowledge of caching and Cache-Control headers? That being said, more and more often in my work I see lots of opportunities being left on the table through unconsidered or even completely overlooked caching practices. Cache-Control. One of the most common and effective ways to manage the caching of your assets is via the Cache-Control HTTP header.

Cache 215

Design Of A Modern Cache—Part Deux

High Scalability

The previous article described the caching algorithms used by Caffeine , in particular the eviction and concurrency models. This allows for quickly discarding new arrivals that are unlikely to be used again, guarding the main region from cache pollution. cache

Cache 230

How to Create a Simple and Efficient PHP Cache


Caching is extremely useful in order to speed up PHP webpages. In this article, I’ll show you how to make a simple PHP caching system for your web pages. tutorial performance cache page caching

Cache 164

Multiple Cache Configurations With Caffeine and Spring Boot


Caching is key for the performance of nearly every application. Distributed caching is sometimes needed, but not always. In many cases, a local cache would work just fine, and there’s no need for the overhead and complexity of the distributed cache. So, in many applications, including plain Spring and Spring Boot, you can use @Cacheable on any method and its result will be cached so that the next time the method is invoked, the cached result is returned.

Cache 100

Using Cache in Spring Boot


A solution to that problem could be using a cache, but how do you implement it? In that article, I explain how to use a basic cache in Spring Boot. java performance spring boot cacheLet's imagine a web application, where for each request received, it must read some configuration data of a database. That data doesn't change usually, but the application, in each request, must connect, execute the correct instructions to read the data, pick it up from the network, etc.

Cache 100

WebP Caching has Landed!


We’re happy to announce that WebP Caching has landed! How Does WebP Caching Work? Enable the Feature for your Zones Cache Key WebP can be enabled for all Pull Zones. Once enabled, a Zone will cache each image separately as WebP and the other image format (e.g.

Cache 99

Design Patterns: Cache-Aside Pattern


Applications that rely heavily on a data-store usually can benefit greatly from using the Cache-Aside Pattern. If used correctly, this pattern can improve performance and help maintain consistency between the cache and the underlying data store. Lifetime of Cached Data.

Cache 64

Exploring a back/forward cache for Chrome

Addy Osmani

bfcache creates a cache allowing for instant navigations to previously-visited pages

Cache 52

Intro to Redis Cluster Sharding – Advantages, Limitations, Deploying & Client Connections

High Scalability

Redis Cluster is the native sharding implementation available within Redis that allows you to automatically distribute your data across multiple nodes without having to rely on external tools and utilities. At ScaleGrid, we recently added support for Redis Clusters on our platform through our fully managed Redis hosting plans.

Cache 188

Compress objects, not cache lines: an object-based compressed memory hierarchy

The Morning Paper

Compress objects, not cache lines: an object-based compressed memory hierarchy Tsai & Sanchez, ASPLOS’19. Existing cache and main memory compression techniques compress data in small fixed-size blocks, typically cache lines.

Cache 66

Re-Architecting the Video Gatekeeper

The Netflix TechBlog

The Tech Hollow , an OSS technology we released a few years ago, has been best described as a total high-density near cache : Total : The entire dataset is cached on each node?—?there there is no eviction policy, and there are no cache misses.

Application Scalability — How To Do Efficient Scaling


performance performance testing performance and monitoring performance tuning performance engineering application scaling performance caching scalability and performance

Self-Host Your Static Assets

CSS Wizardry

Users might already have the file cached. If links to [link] , and a user goes from there to who also links to [link] , then the user will already have that file in their cache. Penalty: Caching. Myth: Cross-Domain Caching.

Cache 284

Expanding the Cloud: More memory, more caching and more performance for your data

All Things Distributed

Amazon ElastiCache is a fully managed, in-memory caching service for customers to optimize the latency, performance and cost of their read workloads.

Cache 84

Redis vs Memcached

Software Architecture

nosql caching memcached database redisMemcached is in-memory key value store whereas Redis is in-memory data structures store. Memcached supports only string data type which is ideal for storing read-only data. Redis supports almost all types of data.

Working with fio “distribution /pereto” parameter


We would expect a similar shape for any sort of caching mechanism. Nutanix X-Ray caching fio modelingThe fio Pareto parameter allows us to create a workload, which references a very large dataset, but specify a hotspot for the access pattern.

Use Parallel Analysis – Not Parallel Query – for Fast Data Access and Scalable Computing Power

ScaleOut Software

Looking beyond distributed caching, it’s their ability to perform data-parallel analysis that gives IMDGs such exciting capabilities. For more than a decade, in-memory data grids (IMDGs) have proven their usefulness for storing fast-changing data in enterprise applications.

Lazy Pre-Browsing with Prefetch

CSS Wizardry

This means that from a cold-cache, if a user were to land on this page for the first time, they’re absolutely going to take a performance hit—there’s just no way around it.

Cache 156

Memory Latency on the Intel Xeon Phi x200 “Knights Landing” processor

John McCalpin

The modes that are important are: “Flat” vs “Cache” In “Flat” mode, MCDRAM memory is used as directly accessible memory, occupying the upper 16 GiB of physical address space. In “Cache” mode, MCDRAM memory is used as an L3 cache for the main DDR4 memory. I will discuss the performance characteristics of Cache mode at a later date.

Why Do We Need the Volatile Keyword?


Even if my application runs in the cloud on the JVM, despite all of those software layers abstracting away the underlying hardware, the volatile keyword is still needed due to the cache of the processor that my software runs on. The Volatile Keyword and the Cache of Modern Processors. Modern processors, like the Intel Xeon or the AMD Ryzen, cache the values from the main memory in per-core caches to improve the memory access performance.

Cache 130

Invited Talk at SuperComputing 2016!

John McCalpin

Computer Architecture Computer Hardware Performance cache DRAM high performance computing memory bandwidth memory latency STREAM benchmark“Memory Bandwidth and System Balance in HPC Systems” If you are planning to attend the SuperComputing 2016 conference in Salt Lake City next month, be sure to reserve a spot on your calendar for my talk on Wednesday afternoon (4:15pm-5:00pm).

Bringing Rich Experiences to Memory-constrained TV Devices

The Netflix TechBlog

Our UI runs on top of a custom rendering engine which uses what we call a “surface cache” to optimize our use of graphics memory. A full-screen image at this resolution will use 1280 * 720 * 4 = 3.5MB of surface cache. The majority of legacy devices run at 28MB of surface cache.

Cache 168

Predictive CPU isolation of containers at Netflix

The Netflix TechBlog

Because microprocessors are so fast, computer architecture design has evolved towards adding various levels of caching between compute units and the main memory, in order to hide the latency of bringing the bits to the brains.

Cache 275

Memory Bandwidth Requirements of the HPL benchmark

John McCalpin

The algorithms used by HPL have lots of data re-use (both in registers and from the caches), but the data still has to go to and from memory, so the bandwidth requirement is not zero, which means that at some point in scaling the number of cores or frequency or FP operations per cycle, we are going to run out of the available memory bandwidth. Algorithms Performance arithmetic cache high performance computing memory bandwidth

Expanding the Cloud - Introducing Amazon ElastiCache - All Things.

All Things Distributed

Today AWS has launched Amazon ElastiCache , a new service that makes it easy to add distributed in-memory caching to any application. Amazon ElastiCache handles the complexity of creating, scaling and managing an in-memory cache to free up brainpower for more differentiating activities.

Cloud 69

Fast key-value stores: an idea whose time has come and gone

The Morning Paper

Generally to cache data (including non-persistent data that never sees a backing store), to share non-persistent data across application services (e.g. Oh, you mean a cache? Yes, a bit like those 2nd-level caches we were talking about earlier, e.g. Ehcache from 2003 onwards.

Cache 97

Making Cloud.typography Fast(er)

CSS Wizardry

To further exacerbate the problem, the 302 response has a Cache-Control: must-revalidate, private. header , meaning that we will always make an outgoing request for this resource regardless of whether or not we’re hitting the site from a cold or a warm cache.

Time to First Byte: What It Is and Why It Matters

CSS Wizardry

only to find that the resource they’re requesting isn’t in that PoP ’s cache.

Speeding up Linux kernel builds with ccache

Nick Desaulniers

ccache , the compiler cache, is a fantastic way to speed up build times for C and C++ code that I previously recommended. Usually when this happens with ccache, there’s something non-deterministic about the builds that prevents cache hits. With ccache, we can check the cache hit/miss stats with -s , clear the cache with -C , and clear the stats with -z. Let’s see what happens to our build time for subsequent builds with a hot cache: No Cache.

Cache 41

Examining the Performance Impact of an Adhoc Workload

SQL Performance

It’s one of the things we look at during a health audit, and Kimberly has a great query from her Plan cache and optimizing for adhoc workloads post that’s part of our toolkit. dm_exec_cached_plans. Once a query plan is in cache, it can be re-used.

Cache 67

MezzFS?—?Mounting object storage in Netflix’s media processing platform

The Netflix TechBlog

Disk Caching? — ? MezzFS can be configured to cache objects on the local disk. Regional caching? —?Netflix If an application in region A is using MezzFS to read from an object stored in region B, MezzFS will cache the object in region A. MezzFS?—?Mounting

Media 276

SQL Server Linux: fsync and Buffered I/O

SQL Server According to Bob

This means data can be stored in file system cache, non-stable media. ) The issue, as described in the link, is that the sync returns the error but may clear the state of the cached pages. Assume the database application opens the backup file, allowing file system caching (~_O_DIRECT.)

CSS and Network Performance

CSS Wizardry

We’re bound to an inefficient caching strategy: a change to, say, the background colour of the currently-selected day on a date picker used on only one page, would require that we cache-bust the entirety of app.css.

Impact of Data locality on DB workloads.


Many different queries are executing in parallel, some hitting RAM cache, some hitting storage. Effect of removing CPU constraints and maintaining data locality on a running DB instance. In this video I migrate a Postgres DB running PGbench benchmark.

View from Nutanix storage during Postgres DB benchmark


Since the DB is small (50% the size of the Linux RAM) – the database is mostly cached on the read side – so we only see writes going to the DB files. A quick look at how the workload is seen from the Nutanix CVM. In this example from prior post.

I Used The Web For A Day On A 50 MB Budget

Smashing Magazine

MB , that suggests I’ve got around 29 pages in my budget, although probably a few more than that if I’m able to stay on the same sites and leverage browser caching. Let’s talk about caching. We’re going to check out Cache-Control. I Used The Web For A Day On A 50 MB Budget.

Cache 97

Benchmarking with Postgres PT2


pgbench with DB size 50% of Linux buffer cache. In this example we run pgbench with a scale factor of 1000 which equates to a database size of around 15GB. The linux VM has 32G RAM, so we don’t expect to see many reads.

Which Query Used the Most CPU? Implementing Extended Events


While you can look at what's in cache through the DMVs to see the queries there, you don't get any real history and you don't get any detail of when the executions occurred. A question that comes up on the forums all the time is, "Which query used the most CPU?"

Cache 130

Time protection: the missing OS abstraction

The Morning Paper

Microarchitectural state of interest includes data and instruction caches, TLBs, branch predictors, instruction- and data-prefetcher state machines, and DRAM row buffers. So these on-core caches must be flushed on a domain switch. Time protection: the missing OS abstraction Ge et al.,

Cache 63

Fostering a Web Performance Culture

José M. Pérez

How would you architecture a non-trivial size web project (client, server, databases, caching layer)? Web Performance is not only about understanding what makes a site fast. It’s about creating awareness amongst both developers and non-developers.

Nutanix AES: Performance By Example PT2


In our experiment we deliberately size the active working-set to NOT fit into the metadata cache. Additionally AES reduces the need to cache metadata in DRAM since local access is so fast. How to improve large DB read performance by 2X. Nutanix AOS 5.10

Cache 52

Service Workers can save the environment!

Dean Hume

Without effective caching on the client, the server will see an increase in workload, more CPU usage and ultimately increased latency for the end user. They allow you to cache resources on the user's device when they visit your site for the first time. Energy Aware Caching.

Optimizing Google Fonts Performance

Smashing Magazine

Browser Caching. Another built-in optimization of Google Fonts is browser caching. As the Google Fonts API becomes more widely used, it is likely visitors to your site or page will already have any Google fonts used in your design in their browser cache.

Google 112