Remove Cache Remove Engineering Remove Traffic
article thumbnail

Consistent caching mechanism in Titus Gateway

The Netflix TechBlog

We introduce a caching mechanism in the API gateway layer, allowing us to offload processing from singleton leader elected controllers without giving up strict data consistency and guarantees clients observe. With traffic growth, a single leader node handling all request volume started becoming overloaded. it will read version E?

Cache 233
article thumbnail

Migrating Netflix to GraphQL Safely

The Netflix TechBlog

By the summer of 2020, many UI engineers were ready to move to GraphQL. The GraphQL shim enabled client engineers to move quickly onto GraphQL, figure out client-side concerns like cache normalization, experiment with different GraphQL clients, and investigate client performance without being blocked by server-side migrations.

Traffic 357
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Predictive CPU isolation of containers at Netflix

The Netflix TechBlog

Because microprocessors are so fast, computer architecture design has evolved towards adding various levels of caching between compute units and the main memory, in order to hide the latency of bringing the bits to the brains. This avoids thrashing caches too much for B and evens out the pressure on the L3 caches of the machine.

Cache 258
article thumbnail

Geek Reading - Week of June 5, 2013

DZone

Improving testing by using real traffic from production ( Hacker News). Using MongoDB as a cache store ( Architects Zone – Architectural Design Patterns & Best Practices). Email Reveals Google App Engine Search API About Ready For Preview Release, Charges Planned For Storage, Operations ( TechCrunch). Hacker News).

Java 244
article thumbnail

Introducing Netflix’s Key-Value Data Abstraction Layer

The Netflix TechBlog

Additionally, the tight coupling with multiple native database APIs — APIs that continually evolve and sometimes introduce backward-incompatible changes — resulted in org-wide engineering efforts to maintain and optimize our microservice’s data access. Either or both may be required by backing storage engines to de-duplicate mutations.

Latency 260
article thumbnail

Cache and Prizes

Alex Russell

Browsers will cache tools popular among vocal, leading-edge developers. There's plenty of space for caching most popular frameworks. The best available proxy data also suggests that shared caches would have a minimal positive effect on performance. Suppose a user has only downloaded part of the cache. None of this holds.

Cache 89
article thumbnail

Seamlessly Swapping the API backend of the Netflix Android app

The Netflix TechBlog

This allowed Android engineers to have much more control and observability over how we get our data. This allows the app to query a list of “paths” in each HTTP request, and get specially formatted JSON (jsonGraph) that we use to cache the data and hydrate the UI. Replay Testing Enter replay testing.

Latency 241