Chashi Mahiul Islam, Alan Villarreal, Mao Nishino, Shaeke Salman, Xiuwen Liu
View original ↗Build a tool to stress-test LLM inference stability across different quantization levels and hardware platforms. Provide metrics for numerical drift and repeatability.
Suggested repo: drift-check
"Is your LLM output actually stable? Find out before you deploy."
Estimated effort: 50h