Remove 2010 Remove Latency Remove Servers
article thumbnail

Single-core memory bandwidth: Latency, Bandwidth, and Concurrency

John McCalpin

The Multicore Era Over the past ~15 years, server processors from Intel and AMD have evolved from the early quad-core processors to the current monsters with over 50 cores per socket. The example below is for a 2005-era processor with 60 ns memory latency and 6.4 If we want to sustain full bandwidth, we need 64/2 =32 cache lines.

Latency 71
article thumbnail

Zero Configuration Service Mesh with On-Demand Cluster Discovery

The Netflix TechBlog

A brief history of IPC at Netflix Netflix was early to the cloud, particularly for large-scale companies: we began the migration in 2008, and by 2010, Netflix streaming was fully run on AWS. In 2010, however, nearly none of it existed: the CNCF wasn’t formed until 2015!

Traffic 229
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

USENIX LISA2021 Computing Performance: On the Horizon

Brendan Gregg

I summarized these topics and more as a plenary conference talk, including my own predictions (as a senior performance engineer) for the future of computing performance, with a focus on back-end servers. Ford, et al., “TCP

article thumbnail

Expanding the Cloud with DNS - Introducing Amazon Route 53 - All.

All Things Distributed

By Werner Vogels on 05 December 2010 02:00 PM. It would not be a first that a customer thinks that his EC2 instance is down when in reality it is some name server somewhere that is not functioning correctly. There are two main types of DNS servers: authoritative servers and caching resolvers. All Things Distributed.

Cloud 111
article thumbnail

Amazon EC2 Cluster GPU Instances - All Things Distributed

All Things Distributed

By Werner Vogels on 14 November 2010 04:00 PM. For example, the most fundamental abstraction trade-off has always been latency versus throughput. The throughput of this pipeline is more important than the latency of the individual operations. No Server Required - Jekyll & Amazon S3. Comments (). Where to go from here?

AWS 125
article thumbnail

Introducing the AWS South America - All Things Distributed

All Things Distributed

This new Region has been highly requested by companies worldwide, and it provides low-latency access to AWS services for those who target customers in South America. The new Sao Paulo Region provides better latency to South America, which enables AWS customers to deliver higher performance services to their South American end-users.

AWS 103
article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP