Remove Design Remove Efficiency Remove Latency
article thumbnail

Netflix’s Distributed Counter Abstraction

The Netflix TechBlog

By: Rajiv Shringi , Oleksii Tkachuk , Kartik Sathyanarayanan Introduction In our previous blog post, we introduced Netflix’s TimeSeries Abstraction , a distributed service designed to store and query large volumes of temporal event data with low millisecond latencies. Today, we’re excited to present the Distributed Counter Abstraction.

Latency 247
article thumbnail

RabbitMQ vs. Kafka: Key Differences

Scalegrid

RabbitMQ is designed for flexible routing and message reliability, while Kafka handles high-throughput event streaming and real-time data processing. Kafka scales efficiently for large data workloads, while RabbitMQ provides strong message durability and precise control over message delivery. What is RabbitMQ?

Latency 147
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Architectural Insights: Designing Efficient Multi-Layered Caching With Instagram Example

DZone

Leveraging this hierarchical structure can significantly reduce latency and improve overall performance.

Cache 173
article thumbnail

Best Practices for Scaling RabbitMQ

Scalegrid

This guide will cover how to distribute workloads across multiple nodes, set up efficient clustering, and implement robust load-balancing techniques. The architecture of RabbitMQ is meticulously designed for complex message routing, enabling dynamic and flexible interactions between producers and consumers.

article thumbnail

API Design Principles for Optimal Performance and Scalability

DZone

It involves a combination of techniques and best practices aimed at reducing latency, improving user experience, and increasing the overall efficiency of the system. API performance optimization is the process of improving the speed, scalability, and reliability of APIs.

article thumbnail

Practical API Design at Netflix, Part 1: Using Protobuf FieldMask

The Netflix TechBlog

Remote calls are never free; they impose extra latency, increase probability of an error, and consume network bandwidth. How can we achieve a similar functionality when designing our gRPC APIs? This (alongside some other techniques like ZigZag encoding for signed types) makes protobuf messages space-efficient.

Design 246
article thumbnail

Timestone: Netflix’s High-Throughput, Low-Latency Priority Queueing System with Built-in Support…

The Netflix TechBlog

Timestone: Netflix’s High-Throughput, Low-Latency Priority Queueing System with Built-in Support for Non-Parallelizable Workloads by Kostas Christidis Introduction Timestone is a high-throughput, low-latency priority queueing system we built in-house to support the needs of Cosmos , our media encoding platform. Over the past 2.5

Latency 217