BenchmarksStack Ranking
APIsPricingDocsWhite PaperTokenBlogAboutSecurity Demo
Log InGet API Key
Performance · 5 min read

67x Proof Cache Speedup:
Intelligent Cryptographic Caching

H33's proof caching system delivers 67x faster verification for returning users. Learn how intelligent caching transforms authentication performance.

2.17M/s
Auth/sec
~42µs
Per Auth
96
CPU Cores
Graviton4
Platform

Cryptographic proof verification is computationally intensive. But most applications repeatedly verify the same types of proofs for the same users. H33's intelligent proof caching eliminates this redundancy, delivering 67x faster verification for cached proofs.

The Cache Advantage

First verification performs full cryptographic validation. Subsequent verifications hit the cache, confirming proof validity in microseconds instead of milliseconds.

The Cost of Repeated Verification

Every authentication event in a post-quantum system requires non-trivial computation. H33's full pipeline performs BFV fully homomorphic encryption over encrypted biometric templates, a ZK proof lookup for identity binding, and a Dilithium signature for attestation. On a single call, this runs in approximately 42 microseconds per user on Graviton4 hardware — fast by any standard. But when the same user authenticates dozens or hundreds of times per session, re-running the full cryptographic stack for every request is wasteful.

Consider a trading platform where a user authenticates at session start, then issues hundreds of API calls per minute. Or a collaborative document editor where each keystroke-sync triggers an access-control check. In these patterns, 80–90% of verification requests come from users whose proofs have already been validated within the last few minutes. Without caching, every one of those requests incurs the full cost of FHE inner product, ZKP validation, and Dilithium verify.

How It Works

When H33 verifies a ZK proof, we don't just validate and discard—we extract a cryptographic fingerprint and cache it:

Why SHA3-256 for Fingerprints

H33 uses SHA3-256 (Keccak) for proof fingerprinting because it is post-quantum secure—no known quantum speedup reduces its collision resistance below 128 bits. Classical hash functions like SHA-256 are also believed quantum-safe, but SHA3 provides domain separation and sponge-based construction that eliminates length-extension attacks entirely.

Security Guarantees

Caching doesn't compromise security. Our cache system maintains full cryptographic guarantees:

// Cache-aware verification
const result = await h33.proof.verify({
  proof: userProof,
  cachePolicy: {
    enabled: true,
    maxAge: '1h',
    contextBinding: ['userId', 'deviceId']
  }
});

// First call: full verification (~42µs)
// Subsequent calls: cache hit (~0.6µs, 67x faster)

Cache Architecture

H33's proof cache is a multi-tier system optimized for different access patterns:

Smart prefetching predicts which proofs will be needed and warms the cache proactively. When a user establishes a session, H33 preloads their recent proof fingerprints into L1 so the first re-verification is already a cache hit.

Why In-Process DashMap Wins at Scale

Early versions of the proof cache used a TCP-based RESP proxy (Redis protocol) for the L1 layer. Under light load, this worked well. But at 96 concurrent workers on Graviton4 metal instances, the single TCP proxy serialized all connections and became a catastrophic bottleneck—throughput collapsed from 1.51M auth/sec to just 136K, an 11x regression.

The solution was replacing the TCP proxy with an in-process DashMap (a concurrent hash map with per-shard locking). The result: 0.085 microsecond lookups with zero contention, even under full 96-worker load. This single architectural change pushed production throughput to 2,172,518 auth/sec—a 5.5% improvement over running raw ZKP verification with no cache at all.

Cache Backend Lookup Latency Auth/sec (96 workers) vs. Baseline
No cache (raw ZKP) 3.7 µs 1,511,624 baseline
TCP RESP Proxy ~50 µs 136,670 -91%
In-process DashMap 0.085 µs 2,172,518 +5.5%

When to Use Caching

Proof caching is most effective for:

On a typical SaaS application with 10,000 daily active users, proof caching reduces total verification CPU time by approximately 85%, because the vast majority of auth checks are repeat verifications within active sessions.

Performance Impact

In typical applications where 80-90% of verifications are for returning users, proof caching delivers dramatic improvements:

Metric Without Cache With Cache (80% hit rate) Improvement
Average latency ~42 µs ~9 µs -78%
P99 latency ~58 µs ~44 µs -24%
Throughput (96 cores) 1.51M/s 1.60M/s +5.5%
CPU utilization 96% ~40% -58%

These aren't theoretical numbers—they're measured across our production infrastructure on c8g.metal-48xl instances running 96 Graviton4 cores. See the full benchmark results for detailed methodology and hardware specifications. The 67x speedup applies to individual cache hits (0.085µs vs. the full ~5.7µs ZKP path). Blended across a realistic workload with 80–90% cache hit rates, the aggregate improvement is what drives the throughput and CPU numbers above.

Integration in One Line

Proof caching requires no configuration changes for most deployments. It is enabled by default on all H33 plans and automatically selects the optimal cache backend based on your infrastructure:

// Single-instance deployment: in-process DashMap (default)
// Multi-container deployment: set CACHEE_MODE=tcp for shared cache
// Custom TTL: override per-request with cachePolicy.maxAge

const auth = await h33.auth.verify({
  biometric: encryptedTemplate,   // BFV FHE encrypted
  proof: zkProof,                 // ZK identity binding
  attestation: dilithiumSig,      // Dilithium signature
  cachePolicy: { maxAge: '5m' }  // optional override
});

For single-instance deployments, H33 defaults to in-process DashMap with zero configuration. For distributed deployments spanning multiple containers, setting CACHEE_MODE=tcp enables a shared Redis-backed cache that synchronizes proof fingerprints across the cluster. The API surface is identical in both modes—only the backend transport changes.

Enable Intelligent Caching

Proof caching is enabled by default on all H33 plans. Experience the speedup yourself.

Get Free API Key

Build With Post-Quantum Security

Enterprise-grade FHE, ZKP, and post-quantum cryptography. One API call. Sub-millisecond latency.

Get Free API Key → Read the Docs
Free tier · 10,000 API calls/month · No credit card required
Verify It Yourself