/u/Just-Ad-6488
View original ↗Build an open-source framework that enables training and inferencing for latent state reasoning models using Mamba architectures. This addresses the high VRAM overhead of traditional Chain-of-Thought approaches by keeping 'reasoning' in the hidden state.
Suggested repo: mamba-think
"Reasoning without the bloat: O(1) memory chain-of-thought using Mamba hidden states."
Estimated effort: 120h