Skip to content

Conversation

@dianed-square
Copy link
Contributor

@dianed-square dianed-square commented Jan 10, 2026

Summary

This PR documents prompt caching which is currently supported by Anthropic, Databricks, OpenRouter, and LiteLLM providers for Claude models

Documentation updates:

  • documentation/docs/getting-started/providers.md:
    • Add tip

Type of Change

  • Feature
  • Bug fix
  • Refactor / Code quality
  • Performance improvement
  • Documentation
  • Tests
  • Security fix
  • Build / Release
  • Other (specify below)

AI Assistance

  • This PR was created or reviewed with AI assistance

Testing

none


Copilot AI review requested due to automatic review settings January 10, 2026 00:48
@dianed-square dianed-square requested a review from a team as a code owner January 10, 2026 00:48
@dianed-square dianed-square changed the title claude prompt caching note docs: claude prompt caching note Jan 10, 2026
| [xAI](https://x.ai/) | Access to xAI's Grok models including grok-3, grok-3-mini, and grok-3-fast with 131,072 token context window. | `XAI_API_KEY`, `XAI_HOST` (optional) |

:::tip Prompt Caching for Claude Models
goose automatically enables Anthropic's [prompt caching](https://platform.claude.com/docs/en/build-with-claude/prompt-caching) when using Claude models via Anthropic, Databricks, OpenRouter, and LiteLLM providers. This adds `cache_control` markers to requests, which can reduce costs for longer conversations by caching frequently-used context. See the [provider implementations](https://github.com/block/goose/tree/main/crates/goose/src/providers) for technical details.
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

linked to providers directory to help future proof the list of providers

Copy link
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

This PR documents prompt caching support for Claude models across multiple providers, improving user awareness of this cost-saving feature.

  • Adds a tip section explaining that goose automatically enables Anthropic's prompt caching for Claude models
  • Clarifies which providers support this feature (Anthropic, Databricks, OpenRouter, and LiteLLM)
  • Provides links to Anthropic's documentation and the implementation code

@github-actions
Copy link
Contributor

github-actions bot commented Jan 10, 2026

PR Preview Action v1.6.3
Preview removed because the pull request was closed.
2026-01-10 01:03 UTC

@dianed-square dianed-square merged commit fa1d874 into main Jan 10, 2026
27 checks passed
@dianed-square dianed-square deleted the docs/claude-prompt-caching branch January 10, 2026 01:00
zanesq added a commit that referenced this pull request Jan 12, 2026
…ased

* 'main' of github.com:block/goose:
  chore: break up process agent response (#6348)
  More 3.7 removal (#6414)
  CLI show extension errors (#6398)
  fix[desktop]: Improve UX for ExtensionItem component (#6443)
  update[doc]: Add tip for GitHub Copilot Provider (#6441)
  Avoid using cliclack.confirm in non-interactive session (#6412)
  docs: claude prompt caching note (#6429)
  Restore task completion notification (#6427)
  docs: stream-json and auth-token cli options (#6426)
fbalicchia pushed a commit to fbalicchia/goose that referenced this pull request Jan 13, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants