article thumbnail

Timestone: Netflix’s High-Throughput, Low-Latency Priority Queueing System with Built-in Support…

The Netflix TechBlog

Timestone: Netflix’s High-Throughput, Low-Latency Priority Queueing System with Built-in Support for Non-Parallelizable Workloads by Kostas Christidis Introduction Timestone is a high-throughput, low-latency priority queueing system we built in-house to support the needs of Cosmos , our media encoding platform. Over the past 2.5

Latency 212
article thumbnail

Stuff The Internet Says On Scalability For July 20th, 2018

High Scalability

That’s mapping applications to the specific architectural choices. The third wing of the architecture piece is the “domain specific system-on-chip.” And you already see that in machine learning, where there’s a really hot field in terms of deep neural nets and other implementations.

Internet 121
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

5 data integration trends that will define the future of ETL in 2018

Abhishek Tiwari

There are several emerging data trends that will define the future of ETL in 2018. In 2018, we anticipate that ETL will either lose relevance or the ETL process will disintegrate and be consumed by new data architectures. Unified data management architecture. Event-driven data flow architecture.

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

New Year’s Updates

John McCalpin

This week I updated three of those posts — two really old ones (primarily of interest to computer architecture historians), and one from 2018: July 2012: Local and Remote Memory Latency on AMD Processors in 2-socket and 4-socket servers. December 2013: Notes on Memory Bandwidth on the Xeon Phi (Knights Corner) Coprocessor.

Latency 40
article thumbnail

Three Other Models of Computer System Performance: Part 1

ACM Sigarch

How many buffers are needed to track pending requests as a function of needed bandwidth and expected latency? Can one both minimize latency and maximize throughput for unscheduled work? These models are useful for insight regarding the basic computer system performance metrics of latency and throughput (bandwidth). Little’s Law.

Systems 60