Remove Architecture Remove Design Remove Latency Remove Tuning
article thumbnail

Why applying chaos engineering to data-intensive applications matters

Dynatrace

Stream processing One approach to such a challenging scenario is stream processing, a computing paradigm and software architectural style for data-intensive software systems that emerged to cope with requirements for near real-time processing of massive amounts of data. We designed experimental scenarios inspired by chaos engineering.

article thumbnail

Allegro Reduces Kafka Producer Latency Outliers by 82% After Switching to XFS

InfoQ

Allegro experimented with different performance optimization options to improve Apache Kafka producer tail latency and eventually switched all its clusters to the XFS filesystem. The company used Kafka protocol sniffing, JVM profiling, and eBPF, which proved instrumental in identifying and eliminating performance bottlenecks.

Latency 111
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

Trending Sources

article thumbnail

LinkedIn Migrates Espresso to HTTP2 and Reduces Connections by 88% and Latency by 75%

InfoQ

to HTTP2, resulting in a reduction in the number of connections, latency, and garbage collection times. LinkedIn was able to dramatically improve the scalability and performance of its Espresso database by migrating it from HTTP1.1 To achieve these gains, the team had to optimize the Netty’s default HTTP2 stack to make it fit their needs.

Latency 99
article thumbnail

MongoDB Best Practices: Security, Data Modeling, & Schema Design

Percona

The main objective of this post is to share my experience over the past years tuning MongoDB and centralize the diverse sources that I crossed in this journey in a unique place. The swap issue is explained in the excellent article by Jeremy Cole at the Swap Insanity and NUMA Architecture. Two other schedulers are deadline and noop.

article thumbnail

Evolving from Rule-based Classifier: Machine Learning Powered Auto Remediation in Netflix Data…

The Netflix TechBlog

Operational automation–including but not limited to, auto diagnosis, auto remediation, auto configuration, auto tuning, auto scaling, auto debugging, and auto testing–is key to the success of modern data platforms. We have also noted a great potential for further improvement by model tuning (see the section of Rollout in Production).

Tuning 210
article thumbnail

How To Scale a Single-Host PostgreSQL Database With Citus

Percona

Rather than listing the concepts, function calls, etc, available in Citus, which frankly is a bit boring, I’m going to explore scaling out a database system starting with a single host. And now, execute the benchmark: -- execute the following on the coordinator node pgbench -c 20 -j 3 -T 60 -P 3 pgbench The results are not pretty.

Database 102
article thumbnail

OEMs: The 6 Questions to Ask Your Data Platform Vendors

VoltDB

If you pick a data platform that can only be deployed in a set number of geographic locations, it could lead to latency issues due to increasingly stringent latency SLAs and trouble meeting those SLAs due to the limits of physics. Early RDBMS products were designed to offer SQL and ACID and not much else.

Latency 52