Keqin Xie
View original ↗Implement a sequence modeling architecture that separates local attention from persistent memory. This addresses the long-context bottleneck without relying on standard KV cache growth.
Suggested repo: lpc-sm-core
"A hybrid architecture that treats long-range context like a persistent memory unit."
Estimated effort: 120h