giuliomagnifico
View original ↗Implement a library that provides traditional, high-performance linguistic benchmarks to challenge LLM outputs. This can serve as a sanity check for AI agents.
Suggested repo: classic-nlp
"Beat the AI at its own game using proven linguistic methods."
Estimated effort: 100h