GitHub - confident-ai/deepeval: The LLM Evaluation Framework

GitHub - confident-ai/deepeval: The LLM Evaluation Framework

github.com
Thumbnail of GitHub - confident-ai/deepeval: The LLM Evaluation Framework

GitHub - BrunoScaglione/langtest: Deliver safe & effective language models

Open LLM Leaderboard - a Hugging Face Space by open-llm-leaderboard

huggingface.co
Thumbnail of Open LLM Leaderboard - a Hugging Face Space by open-llm-leaderboard

Laminar

lmnr.ai
Thumbnail of Laminar

AgentBench: Evaluating LLMs as Agents

arxiv.org