-
Notifications
You must be signed in to change notification settings - Fork 65
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
- Loading branch information
1 parent
3168936
commit 7dcf4a4
Showing
2 changed files
with
6 additions
and
4 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
|
@@ -11,13 +11,15 @@ which includes implementations of many standard metrics. By default, to calculat | |
the `pytrec_eval <https://github.com/cvangysel/pytrec_eval>`_ library, which itself is a Python wrapper around | ||
the widely-used `trec_eval evaluation tool <https://github.com/usnistgov/trec_eval>`_. | ||
|
||
The main way to achieve this is using `pt.Experiment()`. | ||
The main way to achieve this is using ``pt.Experiment()``. If you have an existing results dataframe, you can use | ||
``pt.Evaluate()``. | ||
|
||
API | ||
======== | ||
|
||
.. autofunction:: pyterrier.Experiment() | ||
|
||
.. autofunction:: pyterrier.Evaluate() | ||
|
||
Examples | ||
======== | ||
|
@@ -252,7 +254,7 @@ Often used measures, including the name that must be used, are: | |
- Interpolated recall precision curves (`iprec_at_recall`). This is family of measures, so requesting `iprec_at_recall` will output measurements for `[email protected]`, `[email protected]`, etc. | ||
- Precision at rank cutoff (e.g. `P_5`). | ||
- Recall (`recall`) will generate recall at different cutoffs, such as `recall_5`, etc.). | ||
- Mean response time (`mrt`) will report the average number of milliseconds to conduct a query (this is calculated by `pt.Experiment()` directly, not pytrec_eval). | ||
- Mean response time (`mrt`) will report the average number of milliseconds to conduct a query (this is calculated by ``pt.Experiment()`` directly, not pytrec_eval). | ||
- trec_eval measure *families* such as `official`, `set` and `all_trec` will be expanded. These result in many measures being returned. For instance, asking for `official` results in the following (very wide) output reporting the usual default metrics of trec_eval: | ||
|
||
.. include:: ./_includes/experiment-official.rst | ||
|
@@ -290,7 +292,6 @@ More specifically, lets consider the TREC Deep Learning track passage ranking ta | |
|
||
The available evaluation measure objects are listed below. | ||
|
||
|
||
.. autofunction:: pyterrier.measures.P | ||
|
||
.. autofunction:: pyterrier.measures.R | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters