Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
27 commits
Select commit Hold shift + click to select a range
b8342d1
updated loading in llama demo to use transformer bridge
degenfabian Aug 18, 2025
563c2ef
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Aug 20, 2025
0bc81c1
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Aug 22, 2025
905bba8
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Aug 26, 2025
d65fcc1
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 4, 2025
d472990
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 5, 2025
9406ce0
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 6, 2025
ee20a22
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 7, 2025
c4b7885
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 10, 2025
ef18aaa
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 10, 2025
11ad775
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 12, 2025
fb9e5a2
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 12, 2025
90565c9
Merge remote-tracking branch 'origin/dev-3.x' into llama_demo_transfo…
bryce13950 Sep 12, 2025
878f281
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 10, 2025
ccadc5a
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 13, 2025
a88ec90
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 14, 2025
474d574
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 14, 2025
6e326b6
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 15, 2025
4937aea
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 15, 2025
f7f7e07
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 15, 2025
608f416
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 16, 2025
8a670a1
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 16, 2025
51420cd
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 16, 2025
1624983
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 16, 2025
9cb80fa
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 16, 2025
383d23c
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 17, 2025
1e7af66
Merge remote-tracking branch 'origin/dev-3.x-folding' into llama_demo…
bryce13950 Oct 23, 2025
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions .github/workflows/checks.yml
Original file line number Diff line number Diff line change
Expand Up @@ -230,8 +230,8 @@ jobs:
- "Exploratory_Analysis_Demo"
# - "Grokking_Demo"
# - "Head_Detector_Demo"
# - "Interactive_Neuroscope"
# - "LLaMA"
- "Interactive_Neuroscope"
- "LLaMA"
# - "LLaMA2_GPU_Quantized"
- "Main_Demo"
# - "No_Position_Experiment"
Expand Down
14 changes: 8 additions & 6 deletions demos/LLaMA.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@
},
{
"cell_type": "code",
"execution_count": 4,
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
Expand All @@ -110,7 +110,7 @@
"from transformer_lens.hook_points import (\n",
" HookPoint,\n",
") # Hooking utilities\n",
"from transformer_lens import HookedTransformer\n",
"from transformer_lens.model_bridge import TransformerBridge\n",
"\n",
"torch.set_grad_enabled(False)\n",
"\n",
Expand Down Expand Up @@ -164,7 +164,7 @@
},
{
"cell_type": "code",
"execution_count": 1,
"execution_count": null,
"metadata": {},
"outputs": [],
"source": [
Expand All @@ -174,7 +174,7 @@
" tokenizer = LlamaTokenizer.from_pretrained(MODEL_PATH)\n",
" hf_model = LlamaForCausalLM.from_pretrained(MODEL_PATH, low_cpu_mem_usage=True)\n",
"\n",
" model = HookedTransformer.from_pretrained(\n",
" model = TransformerBridge.boot_transformers(\n",
" \"llama-7b\",\n",
" hf_model=hf_model,\n",
" device=\"cpu\",\n",
Expand All @@ -183,6 +183,7 @@
" center_unembed=False,\n",
" tokenizer=tokenizer,\n",
" )\n",
" model.enable_compatibility_mode()\n",
"\n",
" model = model.to(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n",
" model.generate(\"The capital of Germany is\", max_new_tokens=20, temperature=0)"
Expand All @@ -204,7 +205,7 @@
},
{
"cell_type": "code",
"execution_count": 5,
"execution_count": null,
"metadata": {},
"outputs": [
{
Expand Down Expand Up @@ -274,7 +275,8 @@
"tokenizer = LlamaTokenizer.from_pretrained(LLAMA_2_7B_CHAT_PATH)\n",
"hf_model = LlamaForCausalLM.from_pretrained(LLAMA_2_7B_CHAT_PATH, low_cpu_mem_usage=True)\n",
"\n",
"model = HookedTransformer.from_pretrained(LLAMA_2_7B_CHAT_PATH, device=\"cpu\", fold_ln=False, center_writing_weights=False, center_unembed=False)\n",
"model = TransformerBridge.boot_transformers(LLAMA_2_7B_CHAT_PATH, device=\"cpu\", fold_ln=False, center_writing_weights=False, center_unembed=False)\n",
"model.enable_compatibility_mode()\n",
"\n",
"model = model.to(\"cuda\" if torch.cuda.is_available() else \"cpu\")\n",
"model.generate(\"The capital of Germany is\", max_new_tokens=20, temperature=0)"
Expand Down
Loading