Docs home page
Search...
⌘K
Get started
Overview
Quickstart
Products
Dashboard
Logs
Traces
Prompt management
Evaluations
Gateway
Users
Notifications
Admin
Keywords API keys
LLM provider keys
Team management
Resources
What is LLM monitoring?
Automatic retries
How streaming works
Relari eval
Retrieval
Text generation
Faithfulness
Flesch-Kincaid readability
Answer relevance
Other
Discord
platform
Docs home page
Search...
⌘K
Discord
platform
platform
Search...
Navigation
Text generation
Answer relevance
Documentation
Integrations
API reference
Cookbooks
Changelog ↗
Documentation
Integrations
API reference
Cookbooks
Changelog ↗
Text generation
Answer relevance
Copy page
LLM-based Answer Relevance outputs a score between 0.0 - 1.0 assessing the consistency of the generated answer based on the reference ground truth answers.
Scoring rubric in LLM prompt:
0.0 means that the answer is completely irrelevant to the question.
0.5 means that the answer is partially relevant to the question or it only partially answers the question.
1.0 means that the answer is relevant to the question and completely answers the question.
Settings and parameters
Go to Keywords AI (on top of the left nav bar) > Evaluation > Text generation > Answer Relevance
Click on the
Flesch–Kincaid
card to create the setting:
Click the enable switch to turn on the evaluation
Pick which LLM model you want to use for the evaluation
Make an API call, and the evaluation will be run based on the
Ramdom sampling
setting.
Check results in the
requests log
Was this page helpful?
Yes
No
Suggest edits
Raise issue
Previous
Sentiment analysis
Next
On this page
Scoring rubric in LLM prompt:
Settings and parameters
Assistant
Responses are generated using AI and may contain mistakes.