The LLM Evaluation Framework
-
Updated
Jan 24, 2025 - Python
The LLM Evaluation Framework
LangFair is a Python library for conducting use-case level LLM bias and fairness assessments
[ACL'24] A Knowledge-grounded Interactive Evaluation Framework for Large Language Models
Create an evaluation framework for your LLM based app. Incorporate it into your test suite. Lay the monitoring foundation.
This repo is for an streamlit application that provides a user-friendly interface for evaluating large language models (LLMs) using the beyondllm package.
Add a description, image, and links to the llm-evaluation-metrics topic page so that developers can more easily learn about it.
To associate your repository with the llm-evaluation-metrics topic, visit your repo's landing page and select "manage topics."