HoneyHive AI Docs home page
Search or ask AI...
⌘K
Introduction
Welcome to HoneyHive
Quickstart
Setup & Installation
Troubleshooting & FAQs
Guides
Tracing
Experiments
Monitoring
Overview
Creating Custom Charts
Online Evaluations
Evaluators
Datasets
Prompts
Workspace
Tutorials
Observability Tutorial - RAG
Learn more
Key Concepts
Platform Architecture
Community
Back to Platform
v0.1.x (Stable)
HoneyHive AI Docs home page
v0.1.x (Stable)
Search or ask AI...
⌘K
Ask AI
Community
Back to Platform
Back to Platform
Search...
Navigation
Monitoring
Online Evaluations
Documentation
SDK Reference
API Reference
Integrations
Changelog
Documentation
SDK Reference
API Reference
Integrations
Changelog
Monitoring
Online Evaluations
How to configure online evaluations to monitor your application.
Online evaluations allow you to define domain-specific metrics that can be computed to evaluate your logs asynchronously.
Use encourage using
Sampling
to prevent costs associated with model-graded evaluations at production scale
LLM Evaluators
What
: LLM functions scoring semantic qualities.
Why
: Measure tone, creativity, persuasiveness—things usage metrics miss.
How
:
Create LLM Evaluators
Python Evaluators
What
: Code-defined metrics for precise or complex measurements.
Why
: Compute linguistic metrics, domain-specific scores, etc.
How
:
Create Python Evaluators
LLM Evaluators
Measure the immeasurable with LLM scorers.
Python Evaluators
Ultimate flexibility with custom Python scorers.
Was this page helpful?
Yes
No
Suggest edits
Raise issue
Previous
Introduction
An overview of HoneyHive evaluators
Next
On this page
LLM Evaluators
Python Evaluators
Assistant
Responses are generated using AI and may contain mistakes.