Skip to content

t1128: update model registry with current Anthropic model IDs#1712

Merged
marcusquinn merged 3 commits intomainfrom
feature/t1128
Feb 18, 2026
Merged

t1128: update model registry with current Anthropic model IDs#1712
marcusquinn merged 3 commits intomainfrom
feature/t1128

Conversation

@marcusquinn
Copy link
Owner

@marcusquinn marcusquinn commented Feb 18, 2026

Updates Anthropic model IDs in the model registry to current versions per official docs.

Changes

  • claude-opus-4claude-opus-4-6 (pricing corrected: $15/$75 → $5/$25 per MTok)
  • claude-sonnet-4claude-sonnet-4-6 (pricing unchanged)
  • claude-haiku-3.5claude-haiku-4-5 (pricing updated: $0.80/$4 → $1/$5 per MTok)
  • Updated model-registry.db: both models and subagent_models tables
  • Fixed opus.md frontmatter: was claude-opus-4-20250514, now claude-opus-4-6
  • Fixed haiku.md frontmatter: was claude-3-5-haiku-20241022, now claude-haiku-4-5-20251001
  • Regenerated MODELS.md from updated registry

Verification

Ref #1690

Summary by CodeRabbit

  • Documentation
    • Updated Claude model versions (Opus 4-6, Sonnet 4-6, Haiku 4-5) with new pricing and context specifications.
    • Updated model pricing: Haiku input/output adjusted; Opus costs reduced for improved cost efficiency.
    • Refreshed model documentation and routing recommendations reflecting the latest model releases.
    • Updated performance metrics and task-based model recommendations in the catalog.

- claude-opus-4 -> claude-opus-4-6 ($15/$75 -> $5/$25 per MTok)
- claude-sonnet-4 -> claude-sonnet-4-6 (pricing unchanged)
- claude-haiku-3.5 -> claude-haiku-4-5 ($0.80/$4 -> $1/$5 per MTok)
- Update model-registry.db: models + subagent_models tables
- Fix opus.md: model field was claude-opus-4-20250514, now claude-opus-4-6
- Fix haiku.md: model field was claude-3-5-haiku-20241022, now claude-haiku-4-5-20251001
- Regenerate MODELS.md from updated registry
- Source: https://docs.anthropic.com/en/docs/about-claude/models/overview
@coderabbitai
Copy link
Contributor

coderabbitai bot commented Feb 18, 2026

Warning

Rate limit exceeded

@marcusquinn has exceeded the limit for the number of commits that can be reviewed per hour. Please wait 10 minutes and 28 seconds before requesting another review.

⌛ How to resolve this issue?

After the wait time has elapsed, a review can be triggered using the @coderabbitai review command as a PR comment. Alternatively, push new commits to this PR.

We recommend that you space out your commits to avoid hitting the rate limit.

🚦 How do rate limits work?

CodeRabbit enforces hourly rate limits for each developer per organization.

Our paid plans have higher rate limits than the trial, open-source and free plans. In all cases, we re-allow further reviews after a brief timeout.

Please see our FAQ for further information.

Walkthrough

This change updates Claude model identifiers and associated metadata across the framework from legacy versions (opus-4, sonnet-4, haiku-3.5) to newer releases (opus-4-6, sonnet-4-6, haiku-4-5), refreshing pricing, context limits, cost multipliers, tier mappings, and task recommendations throughout helper scripts and documentation.

Changes

Cohort / File(s) Summary
Helper Scripts
.agents/scripts/compare-models-helper.sh, .agents/scripts/generate-models-md.sh
Updated MODEL_DATA catalog with new Claude identifiers (opus-4-6, sonnet-4-6, haiku-4-5) and refreshed pricing/context values. Modified TIER_MAP and TASK_RECOMMENDATIONS constants to reference new model versions. Adjusted relative cost multipliers: haiku ~0.25x→0.33x, opus ~3x→1.7x. Updated help/scoring text and default recommendations.
Model Documentation
.agents/tools/ai-assistants/models/haiku.md, .agents/tools/ai-assistants/models/opus.md
Updated model identifiers and specs: Haiku to claude-haiku-4-5-20251001 with 64K max output and revised costs ($1.00/$5.00 per 1M); Opus to claude-opus-4-6 with 200K context, 128K output, adjusted costs ($5/$25 per 1M), and ~1.7x cost multiplier.
Models Catalog
MODELS.md
Comprehensive catalog refresh: renamed model identifiers throughout, updated pricing and context data, revised routing/tiers section with new primary models and cost multipliers, refreshed performance leaderboard metrics (successes/totals), and updated task-type distribution reflecting new model lineup.

Estimated code review effort

🎯 3 (Moderate) | ⏱️ ~22 minutes

Possibly related PRs

  • PR #660: Updates the same MODEL_DATA, TIER_MAP, TASK_RECOMMENDATIONS constants in compare-models-helper.sh with new Claude identifiers and pricing.
  • PR #799: Aligns Anthropic Claude model identifier and tier-to-model mappings (opus/sonnet/haiku variants) across the framework.
  • PR #758: Modifies the same model subagent documentation files (haiku.md, opus.md) and tier mappings with refreshed model specifications.

Poem

🔄 Claude's evolution marches on,
New opus-4-6 greets the dawn,
Sonnet and haiku tiers refined,
Costs realigned, performance aligned,
DevOps zen: zero debt, A-grade design ✨

🚥 Pre-merge checks | ✅ 2 | ❌ 1

❌ Failed checks (1 warning)

Check name Status Explanation Resolution
Docstring Coverage ⚠️ Warning Docstring coverage is 25.00% which is insufficient. The required threshold is 80.00%. Write docstrings for the functions missing them to satisfy the coverage threshold.
✅ Passed checks (2 passed)
Check name Status Explanation
Description Check ✅ Passed Check skipped - CodeRabbit’s high-level summary is enabled.
Title check ✅ Passed The title accurately summarizes the main change: updating model registry with current Anthropic model IDs, which is the primary focus across all modified files.

✏️ Tip: You can configure your own custom pre-merge checks in the settings.

✨ Finishing Touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch feature/t1128

Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out.

❤️ Share

Comment @coderabbitai help to get the list of available commands and usage tips.

@gemini-code-assist
Copy link

Summary of Changes

Hello @marcusquinn, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed!

This pull request updates the internal model registry and associated documentation to align with the current Anthropic model IDs and their respective pricing structures. The changes ensure that the system uses the most up-to-date model identifiers and cost information, impacting model recommendations and pricing displays throughout the application. This update improves accuracy and consistency with official Anthropic documentation.

Highlights

  • Anthropic Model ID Updates: Updated claude-opus-4 to claude-opus-4-6, claude-sonnet-4 to claude-sonnet-4-6, and claude-haiku-3.5 to claude-haiku-4-5 across the model registry.
  • Pricing Adjustments: Corrected pricing for claude-opus-4-6 (from $15/$75 to $5/$25 per MTok) and updated claude-haiku-4-5 pricing (from $0.80/$4 to $1/$5 per MTok).
  • Model Metadata Enhancements: Updated frontmatter in opus.md and haiku.md with new model IDs and added 'Max output' details for Opus and Haiku models.
  • Registry and Documentation Regeneration: The internal model database (compare-models-helper.sh) and the public MODELS.md documentation were regenerated to reflect these changes.
Changelog
  • .agents/scripts/compare-models-helper.sh
    • Updated Anthropic model IDs and their pricing in the MODEL_DATA constant.
    • Modified TIER_MAP and TASK_RECOMMENDATIONS to reference the new Anthropic model IDs.
    • Adjusted general model recommendations and scoring examples to use the updated model identifiers.
  • .agents/tools/ai-assistants/models/haiku.md
    • Updated the model ID in the frontmatter.
    • Revised the model details table to reflect the new model ID, updated input/output costs, and added max output information.
  • .agents/tools/ai-assistants/models/opus.md
    • Updated the model ID in the frontmatter.
    • Modified the model details table to include the new model ID, updated context description, input/output costs, and max output information. Also updated the cost comparison note.
  • MODELS.md
    • Updated the 'Last updated' timestamp and pattern data points.
    • Reflected the new Anthropic model IDs and their pricing in the 'Available Models' table.
    • Updated primary model assignments for haiku, sonnet, and opus tiers.
    • Adjusted task counts for sonnet and bugfix categories.
Activity
Using Gemini Code Assist

The full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips.

Invoking Gemini

You can request assistance from Gemini at any point by creating a comment using either /gemini <command> or @gemini-code-assist <command>. Below is a summary of the supported commands on the current page.

Feature Command Description
Code Review /gemini review Performs a code review for the current pull request in its current state.
Pull Request Summary /gemini summary Provides a summary of the current pull request in its current state.
Comment @gemini-code-assist Responds in comments when explicitly tagged, both in pull request comments and review comments.
Help /gemini help Displays a list of available commands.

Customization

To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a .gemini/ folder in the base of the repository. Detailed instructions can be found here.

Limitations & Feedback

Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here.

You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension.

Footnotes

  1. Review the Privacy Notices, Generative AI Prohibited Use Policy, Terms of Service, and learn how to configure Gemini Code Assist in GitHub here. Gemini can make mistakes, so double check it and use code with caution.

@github-actions
Copy link
Contributor

🔍 Code Quality Report

�[0;35m[MONITOR]�[0m Code Review Monitoring Report

�[0;34m[INFO]�[0m Latest Quality Status:
SonarCloud: 0 bugs, 0 vulnerabilities, 24 code smells

�[0;34m[INFO]�[0m Recent monitoring activity:
Wed Feb 18 17:08:22 UTC 2026: Code review monitoring started
Wed Feb 18 17:08:22 UTC 2026: SonarCloud - Bugs: 0, Vulnerabilities: 0, Code Smells: 24

📈 Current Quality Metrics

  • BUGS: 0
  • CODE SMELLS: 24
  • VULNERABILITIES: 0

Generated on: Wed Feb 18 17:08:25 UTC 2026


Generated by AI DevOps Framework Code Review Monitoring

Copy link

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request updates the Anthropic model IDs and their associated pricing across several files, including a helper script and documentation. The changes are consistent with the latest information from Anthropic. However, I've noticed a couple of inconsistencies in the auto-generated MODELS.md file where the relative cost for the updated haiku and opus tiers appear to be stale after the pricing changes. Please see my detailed comments.

MODELS.md Outdated
| Tier | Primary Model | Relative Cost |
|------|---------------|---------------|
| haiku | claude-3-5-haiku | ~0.25x |
| haiku | claude-haiku-4-5 | ~0.25x |

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The relative cost for the haiku tier appears to be outdated. With the new claude-haiku-4-5 pricing ($1/$5 per MTok) and sonnet baseline pricing ($3/$15 per MTok), the cost ratio is approximately 0.33x. The table still shows ~0.25x, which was the ratio for the previous claude-3-5-haiku model.

MODELS.md Outdated
| sonnet | claude-sonnet-4-6 | 1x (baseline) |
| pro | gemini-2.5-pro-preview-06-05 | ~1.5x |
| opus | claude-opus-4 | ~3x |
| opus | claude-opus-4-6 | ~3x |

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

medium

The relative cost for the opus tier seems incorrect. With the new claude-opus-4-6 pricing ($5/$25 per MTok) and sonnet baseline pricing ($3/$15 per MTok), the cost ratio is approximately 1.7x. The table still shows ~3x.

This also conflicts with the opus.md file, which was correctly updated in this PR to mention a 1.7x cost ratio relative to sonnet.

- haiku: ~0.25x -> ~0.33x (claude-haiku-4-5 $1/$5 vs sonnet $3/$15)
- opus: ~3x -> ~1.7x (claude-opus-4-6 $5/$25 vs sonnet $3/$15)
- Regenerate MODELS.md with corrected values
- Addresses Gemini code review feedback on PR #1712
@github-actions
Copy link
Contributor

🔍 Code Quality Report

�[0;35m[MONITOR]�[0m Code Review Monitoring Report

�[0;34m[INFO]�[0m Latest Quality Status:
SonarCloud: 0 bugs, 0 vulnerabilities, 24 code smells

�[0;34m[INFO]�[0m Recent monitoring activity:
Wed Feb 18 17:36:09 UTC 2026: Code review monitoring started
Wed Feb 18 17:36:09 UTC 2026: SonarCloud - Bugs: 0, Vulnerabilities: 0, Code Smells: 24

📈 Current Quality Metrics

  • BUGS: 0
  • CODE SMELLS: 24
  • VULNERABILITIES: 0

Generated on: Wed Feb 18 17:36:12 UTC 2026


Generated by AI DevOps Framework Code Review Monitoring

@github-actions
Copy link
Contributor

🔍 Code Quality Report

�[0;35m[MONITOR]�[0m Code Review Monitoring Report

�[0;34m[INFO]�[0m Latest Quality Status:
SonarCloud: 0 bugs, 0 vulnerabilities, 24 code smells

�[0;34m[INFO]�[0m Recent monitoring activity:
Wed Feb 18 17:39:53 UTC 2026: Code review monitoring started
Wed Feb 18 17:39:53 UTC 2026: SonarCloud - Bugs: 0, Vulnerabilities: 0, Code Smells: 24

📈 Current Quality Metrics

  • BUGS: 0
  • CODE SMELLS: 24
  • VULNERABILITIES: 0

Generated on: Wed Feb 18 17:39:56 UTC 2026


Generated by AI DevOps Framework Code Review Monitoring

@sonarqubecloud
Copy link

Copy link
Contributor

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Caution

Some comments are outside the diff and can’t be posted inline due to platform limitations.

⚠️ Outside diff range comments (2)
.agents/scripts/compare-models-helper.sh (1)

296-298: ⚠️ Potential issue | 🟡 Minor

Stale Last updated dates in three user-facing echo strings.

The code comment on line 146 was correctly updated to 2026-02-18, but the three echo statements that users actually see weren't touched:

  • Line 298: "Prices: USD per 1M tokens. Last updated: 2025-02-08."
  • Line 536: "Last updated: 2025-02-08. Run /compare-models for live pricing check."
  • Line 1025: "Data is embedded in this script. Last updated: 2025-02-08."

These will show an out-of-date timestamp to anyone running list, pricing, or help.

📅 Proposed fix — sync the three displayed dates
-	echo "Prices: USD per 1M tokens. Last updated: 2025-02-08."
+	echo "Prices: USD per 1M tokens. Last updated: 2026-02-18."
-	echo "Last updated: 2025-02-08. Run /compare-models for live pricing check."
+	echo "Last updated: 2026-02-18. Run /compare-models for live pricing check."
-	echo "Data is embedded in this script. Last updated: 2025-02-08."
+	echo "Data is embedded in this script. Last updated: 2026-02-18."
🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In @.agents/scripts/compare-models-helper.sh around lines 296 - 298, Update the
three user-facing echo strings that still show "Last updated: 2025-02-08" to the
correct date "Last updated: 2026-02-18": find and replace the echo statements
containing "Prices: USD per 1M tokens. Last updated: 2025-02-08.", "Last
updated: 2025-02-08. Run /compare-models for live pricing check." and "Data is
embedded in this script. Last updated: 2025-02-08." with the same messages but
using 2026-02-18; ensure only the displayed strings are changed and
formatting/whitespace around the echo calls remains unchanged.
MODELS.md (1)

78-82: ⚠️ Potential issue | 🟡 Minor

Contest results section still references the old model names claude-opus-4 / claude-sonnet-4.

The Quality Scores table was not regenerated against the new model IDs because the scoring DB entries were recorded under the old names. Anyone glancing at MODELS.md will see the catalog claim claude-opus-4-6 while the contest section still names claude-opus-4, creating a confusing split view.

Since MODELS.md is auto-generated from live DBs, the fix must happen at the data layer: the comparisons and responses tables in response-scoring.db need a targeted rename of the old model IDs, or a post-generation sed pass to normalise well-known aliases. A migration script or a note in the PR description would help track this gap.

Would you like me to draft a one-off SQLite migration snippet to rename claude-opus-4claude-opus-4-6 and claude-sonnet-4claude-sonnet-4-6 in the scoring DB, or open a tracking issue for this cleanup?

🤖 Prompt for AI Agents
Verify each finding against the current code and only fix it if needed.

In `@MODELS.md` around lines 78 - 82, The MODELS.md contest table shows outdated
model IDs because the scoring DB recorded entries under old names; run a
targeted data migration to update model IDs in the scoring DB: open
response-scoring.db and UPDATE the comparisons and responses tables replacing
'claude-opus-4' → 'claude-opus-4-6' and 'claude-sonnet-4' → 'claude-sonnet-4-6'
(ensure you update any model_id/model_name columns and related indexes/foreign
keys), or add a post-generation normalisation step that maps known aliases
before writing MODELS.md so generation uses the canonical IDs; include the
migration SQL or alias map in the PR or add a one-off script and document it for
audit.
🤖 Prompt for all review comments with AI agents
Verify each finding against the current code and only fix it if needed.

Outside diff comments:
In @.agents/scripts/compare-models-helper.sh:
- Around line 296-298: Update the three user-facing echo strings that still show
"Last updated: 2025-02-08" to the correct date "Last updated: 2026-02-18": find
and replace the echo statements containing "Prices: USD per 1M tokens. Last
updated: 2025-02-08.", "Last updated: 2025-02-08. Run /compare-models for live
pricing check." and "Data is embedded in this script. Last updated: 2025-02-08."
with the same messages but using 2026-02-18; ensure only the displayed strings
are changed and formatting/whitespace around the echo calls remains unchanged.

In `@MODELS.md`:
- Around line 78-82: The MODELS.md contest table shows outdated model IDs
because the scoring DB recorded entries under old names; run a targeted data
migration to update model IDs in the scoring DB: open response-scoring.db and
UPDATE the comparisons and responses tables replacing 'claude-opus-4' →
'claude-opus-4-6' and 'claude-sonnet-4' → 'claude-sonnet-4-6' (ensure you update
any model_id/model_name columns and related indexes/foreign keys), or add a
post-generation normalisation step that maps known aliases before writing
MODELS.md so generation uses the canonical IDs; include the migration SQL or
alias map in the PR or add a one-off script and document it for audit.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant