diff --git a/docs/docset.yml b/docs/docset.yml index ef39c799f566d..36b1db5604fea 100644 --- a/docs/docset.yml +++ b/docs/docset.yml @@ -20,6 +20,7 @@ subs: ech: "Elastic Cloud Hosted" ess: "Elasticsearch Service" ece: "Elastic Cloud Enterprise" + eck: "Elastic Cloud on Kubernetes" serverless-full: "Elastic Cloud Serverless" security-app: "Elastic Security app" stack-manage-app: "Stack Management" diff --git a/docs/reference/connectors-kibana/elastic-managed-llm.md b/docs/reference/connectors-kibana/elastic-managed-llm.md index f1051b07485e0..aece92f7525ee 100644 --- a/docs/reference/connectors-kibana/elastic-managed-llm.md +++ b/docs/reference/connectors-kibana/elastic-managed-llm.md @@ -1,53 +1,32 @@ --- -navigation_title: "Preconfigured AI Connectors" +navigation_title: "Elastic Managed LLMs" applies_to: stack: ga 9.0 serverless: ga --- -# Preconfigured AI Connectors +# Elastic Managed LLMs -Elastic provides built-in LLMs through AI connectors vetted for GenAI product features across our platform. -Currently, the following built-in LLMs are available: - -* [`anthropic-claude-4.5-sonnet`](#claude-45) -* [`anthropic-claude-3.7-sonnet`](#claude-37) - -## `anthropic-claude-4.5-sonnet` [claude-45] -```{applies_to} -stack: ga 9.3 -serverless: ga -``` - -A high-performance, general-purpose language model suitable for the widest range of use cases. -Model details are available in the [model card](https://assets.anthropic.com/m/12f214efcc2f457a/original/Claude-Sonnet-4-5-System-Card.pdf). - -## `anthropic-claude-3.7-sonnet` [claude-37] -```{applies_to} -stack: ga 9.0 -serverless: ga -``` - -`anthropic-claude-3.7-sonnet` is the default LLM used by the AI connector. - -::::{note} -* The default LLM may change in the future based on evaluations of performance, security, and accuracy. -* This LLM was called **Elastic Managed LLM** before 9.3. -:::: - -Details of the currently used model are available in the [model card](https://assets.anthropic.com/m/785e231869ea8b3b/original/claude-3-7-sonnet-system-card.pdf). +Elastic provides built-in LLMs through managed AI connectors. +These connectors are accessed and managed through the [Elastic {{infer-cap}} Service (EIS)](docs-content://explore-analyze/elastic-inference/eis.md), which is the single entry point for using Elastic Managed LLMs. ## Prerequisites * Requires the `manage_inference` [cluster privilege](https://www.elastic.co/docs/reference/elasticsearch/security-privileges#privileges-list-cluster) (the built-in `inference_admin` role grants this privilege) +* {applies_to}`ece:` {applies_to}`eck:` {applies_to}`self: ga 9.3+` For on-premises installations ({{ece}}, {{eck}}, or self managed clusters), Elastic Managed LLMs are only available through [EIS with Cloud Connect](docs-content://explore-analyze/elastic-inference/connect-self-managed-cluster-to-eis.md). Your {{stack}} version must be 9.3 or later. + +## Available models + +Elastic Managed LLMs are available exclusively through the Elastic {{infer-cap}} Service. +You can find the [list of supported models](docs-content://explore-analyze/elastic-inference/eis.md#supported-models) on the EIS documentation page. ## Region and hosting -The Preconfigured AI Connectors use third party service providers for {{infer}}. Refer to [the Elastic {{infer-cap}} Service page](https://www.elastic.co/docs/explore-analyze/elastic-inference/eis) for details. +The Elastic Managed LLMs use third party service providers for {{infer}}. Refer to [the Elastic {{infer-cap}} Service page](docs-content://explore-analyze/elastic-inference/eis.md) for details. ## Data protection -Customer projects or deployments hosted in any cloud service provider or region have access to the Preconfigured AI Connectors in the AWS US region `us-east-1`. +Customer projects or deployments hosted in any cloud service provider or region have access to Elastic Managed LLMs in the AWS US region `us-east-1`. All data is encrypted in transit. The LLMs are configured for zero data retention: none of the prompts or outputs are stored by the service provider. Only request metadata is logged in AWS CloudWatch. @@ -58,4 +37,4 @@ Read more at our [AI Data FAQs](https://www.elastic.co/trust/ai-data-faq) to lea ## Pricing -The Preconfigured AI Connectors incur a cost per million tokens for input and output tokens. Refer to the Elastic Cloud [pricing pages](https://www.elastic.co/pricing) for details. +Elastic Managed LLMs incur a cost per million tokens for input and output tokens. Refer to the Elastic [pricing page](https://www.elastic.co/pricing) that correspond to your Elastic setup for details.