The Rise of the Internal Cache Platform

How Uber and Mercado Libre scaled distributed caches to billions of ops/sec, and the connection storms, ownership failures, and Valkey decisions that shaped their platforms.
Understanding the NxM Problem in Distributed Caches

When N clients each connect to M nodes, connection load grows multiplicatively. Learn how the NxM problem emerges at scale and why connection storms happen during routine operations.
Why Large Cache Systems Need Routing Layers

Large cache clusters often hit connection and latency limits before memory or CPU. Learn why routing layers and edge caching emerge at scale.
Reduce TTFT by >50% with LMCache + Momento Accelerator

How distributed KV caching with LMCache and Momento Accelerator enables unified access to remote token storage, improving inference efficiency at scale.
Large Objects Ruin the Party – Valkey 9 Tames Them

Tail latencies are where promises break. You can have a system that’s fast 99% of the time, but that 1% is what users remember.
Breakthroughs Are Just Boring Improvements That Pile Up

Engineering wins look dramatic from the outside. From the inside, they’re months of small, unglamorous changes – the kind nobody celebrates until they compound into something that looks impossible.
Cache Rebalancing Was Broken. Here’s How Valkey 9.0 Fixed It

Few things make SREs more nervous than rebalancing a cache cluster. You know the feeling. You add a node, trigger a rebalance, and suddenly latency graphs start jumping. It’s a familiar risk of the job, especially when your cache sits between your users and your database. A small configuration mistake here can suddenly unleash a storm of GET requests on your primary data store. I admit, I never really understood the concept of a slot or why it was needed. But after listening to a recent episode of the Cache It podcast on the new atomic slot migration feature in Valkey 9.0, I finally decided to dig in. The deeper […]
The 5 Metrics that Predict Cache Outages

Learn the 5 critical metrics that predict cache performance issues before they impact users. From P999 latency to cache miss rates – learn what actually matters in production.
Momento Cache – Caching you can trust for data reliability

Ultra-fast caching doesn’t need to be hard. Learn how to assemble the perfect caching strategy to improve data reliability for your app.
MomentoはサーバーレスのRedisではない。ポストRedisの世界のために再構築されたキャッシングです。

Momento Cacheは、ゼロからスケールとシンプルさを追求して作られています。
ラムダ実行コンテキストにおけるgRPCキープアライブの話

Momentoがどのようにしてタイムアウトをなくし、より信頼性の高いキャッシュ・サービスを実現したかをご覧ください。
ElastiCache Serverless has a hidden feature: Memcached replication

Redis is no longer the only way for ElastiCache users to get automatic replication across availability zones.