Remove 2005 Remove Database Remove Storage
article thumbnail

Observability platform vs. observability tools

Dynatrace

Metrics are measures of critical system values, such as CPU utilization or average write latency to persistent storage. For example, in 2005, Dynatrace introduced a distributed tracing tool that allowed developers to implement local tracing and debugging. Observability is made up of three key pillars: metrics, logs, and traces.

article thumbnail

Back-to-Basics Weekend Reading - A Decomposition Storage Model

All Things Distributed

Traditionally records in a database were stored as such: the data in a row was stored together for easy and fast retrieval. Combined with the rise of data warehouse workloads, where there is often significant redundancy in the values stored in columns, and database models based on column oriented storage took off.

Storage 72
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

No Server Required - Jekyll & Amazon S3 - All Things Distributed

All Things Distributed

As some of you may remember I was pretty excited when Amazon Simple Storage Service (S3) released its website feature such that I could serve this weblog completely from S3. I have regenerated all pages since 2005, the pages before that can be found in the "/historical" section. Driving Storage Costs Down for AWS Customers.

Servers 112
article thumbnail

AWS EC2 Virtualization 2017: Introducing Nitro

Brendan Gregg

It's amazing to recall that it was even possible to virtualize x86 before processors had hardware-assisted virtualization (Intel VT-x and AMD-V), which were added in 2005 and 2006. But not all workloads: some are network bound (proxies) and storage bound (databases). ## 5. The AMI and boot are now HVM.

article thumbnail

The Amazing Evolution of In-Memory Computing

ScaleOut Software

In general terms, in-memory computing refers to the related concepts of (a) storing fast-changing data in primary memory instead of in secondary storage and (b) employing scalable computing techniques to distribute a workload across a cluster of servers.

article thumbnail

The Amazing Evolution of In-Memory Computing

ScaleOut Software

In general terms, in-memory computing refers to the related concepts of (a) storing fast-changing data in primary memory instead of in secondary storage and (b) employing scalable computing techniques to distribute a workload across a cluster of servers.

article thumbnail

AWS EC2 Virtualization 2017: Introducing Nitro

Brendan Gregg

It's amazing to recall that it was even possible to virtualize x86 before processors had hardware-assisted virtualization (Intel VT-x and AMD-V), which were added in 2005 and 2006. But not all workloads: some are network bound (proxies) and storage bound (databases). ## 5. The AMI and boot are now HVM.