Cartridges paper
Thread (13 posts)
Can you describe the overall approach of this paper in terms of the framing of "fast weights" (kv cache) and "slow weights" (model parameters)?
The paper frames its approach in terms of a dichotomy between "fast weights" (the KV cache) and "slow weights" (the model parameters):
Key points:
-
Fast weights (KV cache):
- Represented by the Cartridges
- Lightweight, trained per-corpus through self-study
- Capture corpus-specific information
- Enable efficient long-context processing
- Can be composed at inference time
-
Slow weights (model parameters):
- Frozen foundation model
- Provide general language understanding capabilities
- Remain fixed across different corpora
- Serve as the backbone for both ICL and Cartridge processing
The key insight is that by offloading corpus-specific information from the slow weights to the fast weights (Cartridges), they achieve:
- Memory efficiency (38.6x reduction vs ICL)
- Higher throughput (26.4x increase)
- Extended context length (up to 484k tokens)
This approach trades increased offline compute (training Cartridges) for reduced memory usage and higher serving efficiency, making it advantageous in settings where many queries reference the same corpus.
The self-study training process acts as the bridge between the fast and slow weights, allowing the Cartridges to effectively compress corpus information into the KV cache format while leveraging the frozen slow weights' general understanding.