Scorers
Summarization
The Summarization
scorer is a default LLM judge scorer that measures whether your LLM can accurately summarize text.
In this case, the actual_output
is the summary, and the input
is the text to summarize.
Required Fields
To run the Summarization
scorer, you must include the following fields in your Example
:
input
actual_output
Scorer Breakdown
Summarization
scores are calculated by determining:
- Whether the summary contains contradictory information from the original text.
- Whether the summary contains all of the important information from the original text.
To do so, we compute two subscores respectively:
For the information score, we generate a list of important questions from the original text and check the fraction of the questions that are answered by information inthe summary.
The final score is the minimum of the two subscores.
Sample Implementation
summarization.py