Jiuting Chen, Yuan Lian, Hao Wu, Tianqi Huang, Hiroshi Sasaki, Makoto Kouno, Jongil Choi
View original ↗Build a library for probing model uncertainty and internal knowledge boundaries. Developers can use this to detect when a model is hallucinating versus when it genuinely lacks knowledge.
Suggested repo: probeknow
"Does your model actually know? Detect internal ignorance with this probing toolkit."
Estimated effort: 45h