article thumbnail

Optimising for High Latency Environments

CSS Wizardry

This gives fascinating insights into the network topography of our visitors, and how much we might be impacted by high latency regions. Round-trip-time (RTT) is basically a measure of latency—how long did it take to get from one endpoint to another and back again? What is RTT? RTT isn’t a you-thing, it’s a them-thing.

Latency 244
article thumbnail

Comparing Approaches to Durability in Low Latency Messaging Queues

DZone

Little’s Law and Why Latency Matters. In many cases, the assumption is that as long as throughput is high enough, the latency won’t be a problem. However, latency is often a key factor in why the throughput isn’t high enough. This is the first time I have benchmarked it with a realistic example.

Latency 275
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Understanding and Managing Latency in APISIX: A Comprehensive Technical Guide

DZone

A common query from users revolves around the precise measurement of latency in APISIX. When utilizing APISIX, how should one address unusually high latency? In reality, discussions on latency measurement are centered around the performance and response time of API requests.

Latency 261
article thumbnail

Improve Application Latency With Read Replicas Using YugabyteDB [Video]

DZone

Scalability and low latency are crucial for any application that relies on real-time data. In this post, we'll discuss how you can use YugabyteDB and its read replica nodes to improve the read latency for users across the globe. One way to achieve this is by storing data closer to the users.

Latency 246
article thumbnail

Simulate Network Latency and Packet Drop In Linux

DZone

There may be a scenario when you want to test an application when the network is slow(we also call it high network latency). Or you are reproducing a customer scenario(having high network latency) where some anomalous behavior is observed. In the Chrome browser, we can easily Simulate a slower network connection.

Latency 239
article thumbnail

Single-core memory bandwidth: Latency, Bandwidth, and Concurrency

John McCalpin

Latency” is the duration from the execution of a load instruction (to an address that misses in all the caches), and the completion of that load instruction when the data is returned from memory. The example below is for a 2005-era processor with 60 ns memory latency and 6.4 cache lines -> 5.6 cache lines -> 5.6

Latency 71
article thumbnail

Netflix’s Distributed Counter Abstraction

The Netflix TechBlog

By: Rajiv Shringi , Oleksii Tkachuk , Kartik Sathyanarayanan Introduction In our previous blog post, we introduced Netflix’s TimeSeries Abstraction , a distributed service designed to store and query large volumes of temporal event data with low millisecond latencies. Today, we’re excited to present the Distributed Counter Abstraction.

Latency 248