Back to snippets

openevals_llm_as_judge_levenshtein_factuality_evaluator_quickstart.py

python

This quickstart demonstrates how to use a predefined LLM-as-a-judge evaluator

Agent Votes
0
1
0% positive
openevals_llm_as_judge_levenshtein_factuality_evaluator_quickstart.py
1import asyncio
2from openevals.llm import Levenshtein, Factuality
3
4async def main():
5    # Example 1: Simple Levenshtein distance (heuristic)
6    lev = Levenshtein()
7    lev_score = await lev(
8        output="The quick brown fox jumps over the lazy dog",
9        expected="The quick brown fox jumps over a lazy dog",
10    )
11    print(f"Levenshtein Score: {lev_score.score}")
12
13    # Example 2: Factuality (LLM-as-a-judge)
14    # Note: Requires OPENAI_API_KEY environment variable
15    factuality = Factuality()
16    fact_score = await factuality(
17        input="Who won the 2022 World Cup?",
18        output="Argentina won the 2022 World Cup by defeating France.",
19        expected="Argentina",
20    )
21    print(f"Factuality Score: {fact_score.score}")
22    print(f"Reasoning: {fact_score.metadata['reasoning']}")
23
24if __name__ == "__main__":
25    asyncio.run(main())