-
Notifications
You must be signed in to change notification settings - Fork 1.3k
Closed
Labels
Description
Question
I'm running an eval against a dataset with LLMJudge as an evaluator. The table that is printed with this code:
report = dataset.evaluate_sync(get_mongo_query)
report.print(include_input=True, include_output=True)
does not include the evaluation reason explaining why the LLMJudge failed. I see in the LLMJudge code that include_reason=True so wondering if I'm missing some configuration or if there's some other way to see this output. I didn't see any documentation about this.
┏━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━━━━┓
┃ Case ID ┃ Inputs ┃ Outputs ┃ Assertions ┃ Duration ┃
┡━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━━━━┩
│ simple_case │ │ ... │ ✔✗ │ 16.1s │
Additional Context
pydantic_evale v0.2.6
python v3.12.3