article thumbnail

Single-core memory bandwidth: Latency, Bandwidth, and Concurrency

John McCalpin

“Latency” is the duration from the execution of a load instruction (to an address that misses in all the caches), and the completion of that load instruction when the data is returned from memory. GB/s peak DRAM bandwidth, requiring 6 concurrent 64-byte cache line accesses to be pending at all times to maintain full bandwidth.

Latency 71
article thumbnail

Predictive CPU isolation of containers at Netflix

The Netflix TechBlog

Because microprocessors are so fast, computer architecture design has evolved towards adding various levels of caching between compute units and the main memory, in order to hide the latency of bringing the bits to the brains. This avoids thrashing caches too much for B and evens out the pressure on the L3 caches of the machine.

Cache 258
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Kubernetes in the wild report 2023

Dynatrace

On-premises data centers invest in higher capacity servers since they provide more flexibility in the long run, while the procurement price of hardware is only one of many cost factors. Of the organizations in the Kubernetes survey, 71% run databases and caches in Kubernetes, representing a +48% year-over-year increase.

article thumbnail

Distance-Based ISA for Efficient Register Management

ACM Sigarch

To create a CPU core that can execute a large number of instructions in parallel, it is necessary to improve both the architecturewhich includes the overall CPU design and the instruction set architecture (ISA) designand the microarchitecture, which refers to the hardware design that optimizes instruction execution.

article thumbnail

Remote Workstations for the Discerning Artists

The Netflix TechBlog

They need specialized hardware, access to petabytes of images, and digital content creation applications with controlled licenses. Instead, we created a service to take the most popular configurations and cache them. As an engineer, I can work anywhere with a standard laptop as long as I have an IDE and access to Stack Overflow.

article thumbnail

How to Optimize Digital Experience and Operations with Dynatrace

Dynatrace

Reducing CPU Utilization to now only consume 15% of initially provisioned hardware. Missing Cache Settings – Make sure you cache resources that don’t change often on the browser or use a CDN. Missing caching layers, e.g. provide a read-only cache for static data. Missing retry and failover implementations.

Cache 214
article thumbnail

AWS serverless services: Exploring your options

Dynatrace

Instead of worrying about infrastructure management functions, such as capacity provisioning and hardware maintenance, teams can focus on application design, deployment, and delivery. Serverless architecture offers several benefits for enterprises. Simplicity. The first benefit is simplicity. Data Store.