The500Feed.Live
Everything going on in AI - updated daily from 500+ sources
📄 ResearchMay 12, 2026
KV-Fold: One-Step KV-Cache Recurrence for Long-Context Inference
We introduce KV-Fold, a simple, training-free long-context inference protocol that treats the key-value (KV) cache as the accumulator in a left fold over sequence chunks. At each step, the model processes the next chunk conditioned on the accumulated cache, appends the newly produced keys and values...
Read Original Article →