diff --git a/README.md b/README.md
index ba5f698b8..d5200f557 100644
--- a/README.md
+++ b/README.md
@@ -25,7 +25,7 @@
-
+
@@ -44,7 +44,7 @@ sample-by-sample results* to debug and see how your models stack-up.
Lighteval supports **1000+ evaluation tasks** across multiple domains and
languages. Use [this
-space](https://huggingface.co/spaces/SaylorTwift/benchmark_finder) to find what
+space](https://huggingface.co/spaces/OpenEvals/open_benchmark_index) to find what
you need, or, here's an overview of some *popular benchmarks*:
@@ -107,6 +107,7 @@ huggingface-cli login
Lighteval offers the following entry points for model evaluation:
+- `lighteval eval`: Evaluation models using [inspect-ai](https://inspect.aisi.org.uk/) as a backend (prefered).
- `lighteval accelerate`: Evaluate models on CPU or one or more GPUs using [🤗
Accelerate](https://github.com/huggingface/accelerate)
- `lighteval nanotron`: Evaluate models in distributed settings using [⚡️
@@ -126,9 +127,7 @@ Did not find what you need ? You can always make your custom model API by follow
Here's a **quick command** to evaluate using the *Accelerate backend*:
```shell
-lighteval accelerate \
- "model_name=gpt2" \
- "leaderboard|truthfulqa:mc|0"
+lighteval eval "hf-inference-providers/openai/gpt-oss-20b" "lighteval|gpqa:diamond|0"
```
Or use the **Python API** to run a model *already loaded in memory*!
diff --git a/docs/source/_toctree.yml b/docs/source/_toctree.yml
index d3b9c9d9b..d3c33cdab 100644
--- a/docs/source/_toctree.yml
+++ b/docs/source/_toctree.yml
@@ -7,6 +7,8 @@
title: Quicktour
title: Getting started
- sections:
+ - local: inspect-ai
+ title: Examples using Inspect-AI
- local: saving-and-reading-results
title: Save and read results
- local: caching
diff --git a/docs/source/available-tasks.mdx b/docs/source/available-tasks.mdx
index 450b7ed49..57605577a 100644
--- a/docs/source/available-tasks.mdx
+++ b/docs/source/available-tasks.mdx
@@ -1,6 +1,8 @@
+# Available tasks
+Browse and inspect tasks available in LightEval.