evidently - An open-source framework to evaluate, test and monitor ML and LLM-powered systems. #877
Labels
llm
Large Language Models
llm-evaluation
Evaluating Large Language Models performance and behavior through human-written evaluation sets
MachineLearning
ML Models, Training and Inference
Observability
Instrumentation and observation of running systems
Software2.0
Software development driven by AI and neural networks.
source-code
Code snippets
technical-writing
Links to deep technical writing and books
use-cases
user use case descriptions
evidently/README.md at main · evidentlyai/evidently
Evidently
An open-source framework to evaluate, test and monitor ML and LLM-powered systems.
Documentation | Discord Community | Blog | Twitter | Evidently Cloud
🆕 New release
Evidently 0.4.25. LLM evaluation -> Tutorial
📊 What is Evidently?
Evidently is an open-source Python library for ML and LLM evaluation and observability. It helps evaluate, test, and monitor AI-powered systems and data pipelines from experimentation to production.
Evidently is very modular. You can start with one-off evaluations using
Reports
orTest Suites
in Python or get a real-time monitoringDashboard
service.1. Reports
Reports compute various data, ML and LLM quality metrics. You can start with Presets or customize.
2. Test Suites
Test Suites check for defined conditions on metric values and return a pass or fail result.
gt
(greater than),lt
(less than), etc.3. Monitoring Dashboard
Monitoring UI service helps visualize metrics and test results over time.
You can choose:
Evidently Cloud offers a generous free tier and extra features like user management, alerting, and no-code evals.
👩💻 Install Evidently
Evidently is available as a PyPI package. To install it using pip package manager, run:
To install Evidently using conda installer, run:
Option 1: Test Suites
Import the Test Suite, evaluation Preset and toy tabular dataset.
Split the
DataFrame
into reference and current. Run the Data Stability Test Suite that will automatically generate checks on column value ranges, missing values, etc. from the reference. Get the output in Jupyter notebook:You can also save an HTML file. You'll need to open it from the destination folder.
To get the output as JSON:
You can choose other Presets, individual Tests and set conditions.
Option 2: Reports
Import the Report, evaluation Preset and toy tabular dataset.
Run the Data Drift Report that will compare column distributions between
current
andreference
:Save the report as HTML. You'll later need to open it from the destination folder.
To get the output as JSON:
You can choose other Presets and individual Metrics, including LLM evaluations for text data.
Option 3: ML monitoring dashboard
Recommended step: create a virtual environment and activate it.
After installing Evidently (
pip install evidently
), run the Evidently UI with the demo projects:Access Evidently UI service in your browser. Go to the localhost:8000.
💻 Contributions
We welcome contributions! Read the Guide to learn more.
📚 Documentation
For more information, refer to a complete Documentation. You can start with the tutorials:
See more examples in the Docs.
How-to guides
Explore the How-to guides to understand specific features in Evidently.
✅ Discord Community
If you want to chat and connect, join our Discord community.
Suggested labels
None
The text was updated successfully, but these errors were encountered: