openai-evals
NousResearch/openai-evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
1stars
Forks
0
Open issues
0
Watchers
1
Size
6.3 MB
Other
Created: Nov 29, 2025
Updated: Apr 3, 2026
Last push: Nov 29, 2025
Fork