Remove Demo Remove Hardware Remove Latency
article thumbnail

What is serverless computing? Driving efficiency without sacrificing observability

Dynatrace

This allows teams to sidestep much of the cost and time associated with managing hardware, platforms, and operating systems on-premises, while also gaining the flexibility to scale rapidly and efficiently. When an application is triggered, it can cause latency as the application starts. This creates latency when they need to restart.

article thumbnail

USENIX LISA2021 Computing Performance: On the Horizon

Brendan Gregg

This was a chance to talk about other things I've been working on, such as the present and future of hardware performance. The video is on [youtube]: The slides are on [slideshare] or as a [PDF]: I work on many areas of performance, but recently I've had a lot of demand to talk about BPF. Ford, et al., “TCP

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP

article thumbnail

Monitoring Distributed Systems

Dotcom-Montior

Software and hardware components are autonomous and execute tasks concurrently. A distributed system comprises of a variety of hardware and software components with different operating systems and technologies, meaning the processors are separate and independent of each other. State is distributed through the system. Concurrency.

Systems 74
article thumbnail

Volt Significantly Faster, and Cheaper, than Intel on (AWS) ARM

VoltDB

The raw performance improvement is due to c7g not using hardware threading, while Intel uses two hardware threads per physical core, which increases latency once you get beyond 50% CPU usage. We then moved up to a more powerful 3-node cluster. While this could be addressed, it was outside the scope of this benchmark.

AWS 52
article thumbnail

USENIX LISA2021 Computing Performance: On the Horizon

Brendan Gregg

This was a chance to talk about other things I've been working on, such as the present and future of hardware performance. The video is on [youtube]: The slides are [here] or as a [PDF]: first prev next last / permalink/zoom I work on many areas of performance, but recently I've had a lot of demand to talk about BPF. Ford, et al., “TCP

article thumbnail

USENIX SREcon APAC 2022: Computing Performance: What's on the Horizon

Brendan Gregg

My personal opinion is that I don't see a widespread need for more capacity given horizontal scaling and servers that can already exceed 1 Tbyte of DRAM; bandwidth is also helpful, but I'd be concerned about the increased latency for adding a hop to more memory. Ford, et al., “TCP