Momento is heading to QCon SF! Meet us at Valkey Booth #32. Learn more ➜

Valkey 9.0 – The Next Generation of Caching

カワジャ・シャムス ヘッドショット カワジャ・シャムス
アレン・ヘルトン

When you build high-stakes systems where milliseconds matter and downtime isn’t an option, every architectural detail counts. Valkey 9.0 represents the most robust caching release we’ve seen in years. The new features take major leaps forward on the fundamentals that matter: resilience, developer experience, and raw speed.

Whether you’re building globally-available caches, hyperscale viewer experiences, or running mission-critical latency-sensitive workloads like gaming platforms, trading engines, and media origins, this release deserves your attention.

Resilience through atomic slot migration

Let’s start with what matters most in production – reliability under load.

When you scale a cluster under load, rebalancing can quietly become one of the riskiest operations you’ll run. You’re moving live data while traffic keeps flowing, keeping in-flight commands consistent, routing new keys correctly, and surviving node transitions without dropping requests.

Valkey 9.0 introduces atomic slot migration, a long-awaited upgrade to how clusters rebalance data. Instead of stepwise migrations where ownership changes mid-flight, slot movement is now handled atomically, guaranteeing consistent routing for new keys and predictable hand-off for existing ones. This minimizes transient errors and reduces the manual choreography that used to come with live resharding.

Under the hood, Valkey coordinates migration states through the CLUSTER SETSLOT API and cluster metadata replication, keeping primaries and replicas aligned on slot state. Now you get safer, faster cluster operations. Atomic slot migration helps you scale faster, safer, with less manual work and more confidence.

For teams running Valkey in clustered environments, this fundamentally shifts how you plan capacity and manage operational risk. Scale-outs become predictable instead of painful.

DX improvements with hash-field expirations

Great infrastructure goes beyond performance. It lets you model data the way it actually works.

You know the pain – a hash with 20 fields, some need to expire in seconds, others persist for hours or days. Until now, you had to manage each as a separate key or embed timestamps and handle expiry logic in your app.

With Valkey 9.0, you can now assign per-field TTLs inside a hash. Commands like HEXPIRE, HPEXPIRE, HEXPIREAT, and HPEXPIREAT let you set or modify expirations on specific fields, while HTTL, HPTTL, and their variants let you query them. The implementation integrates with Valkey’s existing expiry system: once a field’s timer runs out, it’s treated as nonexistent by all hash commands and automatically cleaned up by a background job.

This opens up new patterns for modeling data. Now, a user’s temporary session token can expire in 60 seconds while their preferences stay forever. Or game state can continually refresh while static leaderboard metadata persists across rounds. Or per-item relevance windows in a link graph or feed cache can self-expire without key fragmentation.

Per-field TTLs mean fewer keys, simpler logic, and cleaner memory behavior. You no longer need to explode a single logical entity into dozens of keys just to get the expiration semantics right. The result is a cache that mirrors your data’s lifecycle, not one that forces your data model to adapt to the cache.

It’s a small change that simplifies whole classes of architectural patterns.

Speed boosts SIMD-Accelerated Performance

Resilience and developer experience set the foundation. Speed makes it sing.

The performance story in 9.0 is about using modern CPU capabilities intelligently. Valkey 9.0 takes advantage of modern hardware to deliver performance that outpaces previous versions by wide margins.

First, Valkey brings AVX-512 acceleration to string-to-integer conversions, resulting in a ~19% improvement. It’s a subtle optimization with a huge ripple effect: every INCR, every sorted-set score, every TTL calculation depends on this conversion path. When you’re processing millions of commands per second, a faster parser directly translates to higher throughput and flatter tail latency.

Next, BITCOUNT has been supercharged with AVX2 SIMD vectorization, showing gains from 10% at 4 KB value sizes to more than 500% at 1 MB! For workloads built on analytics, counters, and bitfield tracking, this means entire CPU cores freed up. Now, real-time telemetry, feature flagging, and session tracking workloads can operate at a fraction of their previous cost.

Your infrastructure literally does more with the same hardware.

Add ARM NEON and SSE2 optimizations for hash-table lookups (up to 6% faster), combine them with smarter pipeline parsing and multi-command prefetching, and the result is a system that keeps every core busy doing useful work.

As a result, you get lower tail latency, higher per-node throughput, and measurable cost efficiency at scale – the kind of performance uplift you can actually feel in production.

Community-powered innovation

In the short time since its founding in March 2024, the Valkey project has already shipped two major releases and expanded from 22 to 47 corporate participants under the Linux Foundation’s open governance model. That kind of growth reflects a community aligned on what infrastructure should be: reliable, transparent, and built for the long haul.

What you see in Valkey 9.0 isn’t the work of a single team pushing their roadmap. It’s the result of diverse operators, developers, and engineers solving real problems together. Atomic slot migration exists because operators lived through painful resharding incidents. Hash-field expirations emerged from developers who were tired of fragmenting their data models just to get TTL semantics right. SIMD optimizations came from teams squeezing every cycle out of their hardware at scale.

Each feature reflects a different kind of robustness: operational safety when you’re moving data under load, cleaner code when you’re modeling complex state, and headroom when you’re pushing throughput limits. Together, they represent a step change in what’s possible with open-source caching.

The ecosystem is moving just as fast. New module SDKs, including a Rust-based toolkit maintained by the community, are expanding what you can build safely. Support for document-oriented operations (JSON) and probabilistic structures (Bloom filters) points to a richer, more composable caching layer ahead.

This is what open governance looks like in practice: a system that evolves to meet the needs of the people actually running it in production. If you’re already using Valkey, or thinking about it, this is the release to test. Push it under load, try out atomic slot migration during a scale-out, and revisit those places where you’ve been splitting hashes just to manage per-field TTLs.

Valkey 9.0 raises the bar for what caching infrastructure can be. Predictable resilience at scale. Developer experience that matches how data actually lives and dies. Performance that keeps getting better as hardware evolves. And a community that ensures these improvements keep coming, release after release.

At Momento, we specialize in high-performance Valkey for web-scale applications. Release after release, we bring the latest improvements in performance and safety to our customers without any operational overhead. If you want to see what modern caching feels like at scale, start building with Momento.

On this page