Remove Cache Remove Hardware Remove Latency Remove Transportation
article thumbnail

Why I hate MPI (from a performance analysis perspective)

John McCalpin

According to Dr. Bandwidth, performance analysis has two recurring themes: How fast should this code (or “simple” variations on this code) run on this hardware? The user environment defines the mapping of MPI ranks to hardware resources (cores, sockets, nodes). The MPI runtime library. in ways that are seldom transparent.

article thumbnail

Front-End Performance Checklist 2021

Smashing Magazine

Build Optimizations JavaScript modules, module/nomodule pattern, tree-shaking, code-splitting, scope-hoisting, Webpack, differential serving, web worker, WebAssembly, JavaScript bundles, React, SPA, partial hydration, import on interaction, 3rd-parties, cache. This improves page load time and caching during navigations.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

Front-End Performance Checklist 2020 [PDF, Apple Pages, MS Word]

Smashing Magazine

Estimated Input Latency tells us if we are hitting that threshold, and ideally, it should be below 50ms. Designed for the modern web, it responds to actual congestion, rather than packet loss like TCP does, it is significantly faster , with higher throughput and lower latency — and the algorithm works differently.

article thumbnail

Front-End Performance Checklist 2019 [PDF, Apple Pages, MS Word]

Smashing Magazine

Estimated Input Latency tells us if we are hitting that threshold, and ideally, it should be below 50ms. On the other hand, we have hardware constraints on memory and CPU due to JavaScript parsing times (we’ll talk about them in detail later). A good starting point is to choose a good default stack for your application.

article thumbnail

Can You Afford It?: Real-world Web Performance Budgets

Alex Russell

Contended, over-subscribed cells can make “fast” networks brutally slow, transport variance can make TCP much less efficient , and the bursty nature of web traffic works against us. It simulates a link with a 400ms RTT and 400-600Kbps of throughput (plus latency variability and simulated packet loss).

article thumbnail

HTTP/3: Performance Improvements (Part 2)

Smashing Magazine

Because we are dealing with network protocols here, we will mainly look at network aspects, of which two are most important: latency and bandwidth. Latency can be roughly defined as the time it takes to send a packet from point A (say, the client) to point B (the server). Two-way latency is often called round-trip time (RTT).