-
Notifications
You must be signed in to change notification settings - Fork 153
Evals docs in Logfire (points to Pydantic AI) #1374
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Open
summerscope
wants to merge
10
commits into
main
Choose a base branch
from
laura/evals-docs
base: main
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
+92
−0
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Deploying logfire-docs with
|
Latest commit: |
6bf3134
|
Status: | ✅ Deploy successful! |
Preview URL: | https://a1fc6e0f.logfire-docs.pages.dev |
Branch Preview URL: | https://laura-evals-docs.logfire-docs.pages.dev |
summerscope
commented
Sep 10, 2025
dmontagu
reviewed
Sep 10, 2025
|
||
!!! note "Code-First Evaluation" | ||
|
||
Evals are created and run using the [Pydantic Evals](https://ai.pydantic.dev/evals/) a sub-package of Pydantic AI. Logfire serves as a read-only observability layer where you can view and compare results. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Suggested change
Evals are created and run using the [Pydantic Evals](https://ai.pydantic.dev/evals/) a sub-package of Pydantic AI. Logfire serves as a read-only observability layer where you can view and compare results. | |
Evals are created and run using the [Pydantic Evals](https://ai.pydantic.dev/evals/) package, which is developed in tandem with Pydantic AI. Logfire serves as an observability layer where you can view and compare results. |
dmontagu
reviewed
Sep 10, 2025
@@ -0,0 +1,91 @@ | |||
# Evals (beta) | |||
|
|||
View and analyze your evaluation results in Pydantic Logfire's web interface. Evals provides observability into how your AI systems perform across different test cases and experiments over time. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Suggested change
View and analyze your evaluation results in Pydantic Logfire's web interface. Evals provides observability into how your AI systems perform across different test cases and experiments over time. | |
View and analyze your evaluation results in Pydantic Logfire's web interface. Evals provide observability into how your AI systems perform across different test cases and experiments over time. |
dmontagu
reviewed
Sep 10, 2025
|
||
## What are Evals? | ||
|
||
Evals help you systematically test and evaluate AI systems by running them against predefined test cases. Each evaluation experiment appears in Logfire automatically when you run the Pydantic Evals package with Logfire integration enabled. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Suggested change
Evals help you systematically test and evaluate AI systems by running them against predefined test cases. Each evaluation experiment appears in Logfire automatically when you run the Pydantic Evals package with Logfire integration enabled. | |
Evals help you systematically test and evaluate AI systems by running them against predefined test cases. Each evaluation experiment appears in Logfire automatically when you run the `pydantic_evals.Dataset.evaluate` method with Logfire integration enabled. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This isn't precious - just trying to get something into Logfire docs to start with, feel free to slash / edit as you feel fit. Mostly just wanted to make sure people go to the Pydantic AI docs to read about Evals properly.