Skip to content

Conversation

summerscope
Copy link
Contributor

@summerscope summerscope commented Sep 5, 2025

This isn't precious - just trying to get something into Logfire docs to start with, feel free to slash / edit as you feel fit. Mostly just wanted to make sure people go to the Pydantic AI docs to read about Evals properly.

Copy link

cloudflare-workers-and-pages bot commented Sep 5, 2025

Deploying logfire-docs with  Cloudflare Pages  Cloudflare Pages

Latest commit: 6bf3134
Status: ✅  Deploy successful!
Preview URL: https://a1fc6e0f.logfire-docs.pages.dev
Branch Preview URL: https://laura-evals-docs.logfire-docs.pages.dev

View logs

@summerscope summerscope marked this pull request as ready for review September 10, 2025 08:58
@summerscope summerscope changed the title Evals docs Evals docs in Logfire (points to Pydantic AI) Sep 10, 2025

!!! note "Code-First Evaluation"

Evals are created and run using the [Pydantic Evals](https://ai.pydantic.dev/evals/) a sub-package of Pydantic AI. Logfire serves as a read-only observability layer where you can view and compare results.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Evals are created and run using the [Pydantic Evals](https://ai.pydantic.dev/evals/) a sub-package of Pydantic AI. Logfire serves as a read-only observability layer where you can view and compare results.
Evals are created and run using the [Pydantic Evals](https://ai.pydantic.dev/evals/) package, which is developed in tandem with Pydantic AI. Logfire serves as an observability layer where you can view and compare results.

@@ -0,0 +1,91 @@
# Evals (beta)

View and analyze your evaluation results in Pydantic Logfire's web interface. Evals provides observability into how your AI systems perform across different test cases and experiments over time.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
View and analyze your evaluation results in Pydantic Logfire's web interface. Evals provides observability into how your AI systems perform across different test cases and experiments over time.
View and analyze your evaluation results in Pydantic Logfire's web interface. Evals provide observability into how your AI systems perform across different test cases and experiments over time.


## What are Evals?

Evals help you systematically test and evaluate AI systems by running them against predefined test cases. Each evaluation experiment appears in Logfire automatically when you run the Pydantic Evals package with Logfire integration enabled.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Suggested change
Evals help you systematically test and evaluate AI systems by running them against predefined test cases. Each evaluation experiment appears in Logfire automatically when you run the Pydantic Evals package with Logfire integration enabled.
Evals help you systematically test and evaluate AI systems by running them against predefined test cases. Each evaluation experiment appears in Logfire automatically when you run the `pydantic_evals.Dataset.evaluate` method with Logfire integration enabled.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants