Skip to content

See eval reasons in report printed report #1809

@brian-swantide

Description

@brian-swantide

Question

I'm running an eval against a dataset with LLMJudge as an evaluator. The table that is printed with this code:

 report = dataset.evaluate_sync(get_mongo_query)
 report.print(include_input=True, include_output=True)

does not include the evaluation reason explaining why the LLMJudge failed. I see in the LLMJudge code that include_reason=True so wondering if I'm missing some configuration or if there's some other way to see this output. I didn't see any documentation about this.

┏━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━━━━━━━━┳━━━━━━━━━━━━┳━━━━━━━━━━┓
┃ Case ID     ┃ Inputs             ┃ Outputs           ┃ Assertions ┃ Duration ┃
┡━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━━━━━━━━╇━━━━━━━━━━━━╇━━━━━━━━━━┩
│ simple_case │                    │ ...     │ ✔✗         │    16.1s │

Additional Context

pydantic_evale v0.2.6
python v3.12.3

Metadata

Metadata

Assignees

Labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions