Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 8 additions & 8 deletions config/serverless.es.yml
Original file line number Diff line number Diff line change
Expand Up @@ -116,8 +116,8 @@ xpack.contentConnectors.ui.enabled: false

# Elastic Managed LLMs
xpack.actions.preconfigured:
General-Purpose-LLM-v1:
name: General Purpose LLM v1
Anthropic-Claude-Sonnet-3-7:
name: Anthropic Claude Sonnet 3.7
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -126,8 +126,8 @@ xpack.actions.preconfigured:
inferenceId: ".rainbow-sprinkles-elastic"
providerConfig:
model_id: "rainbow-sprinkles"
General-Purpose-LLM-v2:
name: General Purpose LLM v2
Anthropic-Claude-Sonnet-4-5:
name: Anthropic Claude Sonnet 4.5
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -136,13 +136,13 @@ xpack.actions.preconfigured:
inferenceId: ".gp-llm-v2-chat_completion"
providerConfig:
model_id: "gp-llm-v2"
General-Purpose-LLM-v3:
name: General Purpose LLM v3
OpenAI-GPT-OSS-120B:
name: OpenAI GPT-OSS 120B
actionTypeId: .inference
exposeConfig: true
config:
provider: "elastic"
taskType: "chat_completion"
inferenceId: ".gp-llm-v3-chat_completion"
inferenceId: ".openai-gpt-oss-120b-chat_completion"
providerConfig:
model_id: "gp-llm-v3"
model_id: "openai-gpt-oss-120b"
17 changes: 9 additions & 8 deletions config/serverless.oblt.complete.yml
Original file line number Diff line number Diff line change
Expand Up @@ -25,8 +25,8 @@ xpack.features.overrides:

# Elastic Managed LLMs
xpack.actions.preconfigured:
General-Purpose-LLM-v1:
name: General Purpose LLM v1
Anthropic-Claude-Sonnet-3-7:
name: Anthropic Claude Sonnet 3.7
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -35,8 +35,8 @@ xpack.actions.preconfigured:
inferenceId: ".rainbow-sprinkles-elastic"
providerConfig:
model_id: "rainbow-sprinkles"
General-Purpose-LLM-v2:
name: General Purpose LLM v2
Anthropic-Claude-Sonnet-4-5:
name: Anthropic Claude Sonnet 4.5
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -45,13 +45,14 @@ xpack.actions.preconfigured:
inferenceId: ".gp-llm-v2-chat_completion"
providerConfig:
model_id: "gp-llm-v2"
General-Purpose-LLM-v3:
name: General Purpose LLM v3
OpenAI-GPT-OSS-120B:
name: OpenAI GPT-OSS 120B
actionTypeId: .inference
exposeConfig: true
config:
provider: "elastic"
taskType: "chat_completion"
inferenceId: ".gp-llm-v3-chat_completion"
inferenceId: ".openai-gpt-oss-120b-chat_completion"
providerConfig:
model_id: "gp-llm-v3"
model_id: "openai-gpt-oss-120b"

16 changes: 8 additions & 8 deletions config/serverless.security.complete.yml
Original file line number Diff line number Diff line change
Expand Up @@ -9,8 +9,8 @@ xpack.features.overrides:

# Elastic Managed LLMs
xpack.actions.preconfigured:
General-Purpose-LLM-v1:
name: General Purpose LLM v1
Anthropic-Claude-Sonnet-3-7:
name: Anthropic Claude Sonnet 3.7
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -19,8 +19,8 @@ xpack.actions.preconfigured:
inferenceId: ".rainbow-sprinkles-elastic"
providerConfig:
model_id: "rainbow-sprinkles"
General-Purpose-LLM-v2:
name: General Purpose LLM v2
Anthropic-Claude-Sonnet-4-5:
name: Anthropic Claude Sonnet 4.5
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -29,13 +29,13 @@ xpack.actions.preconfigured:
inferenceId: ".gp-llm-v2-chat_completion"
providerConfig:
model_id: "gp-llm-v2"
General-Purpose-LLM-v3:
name: General Purpose LLM v3
OpenAI-GPT-OSS-120B:
name: OpenAI GPT-OSS 120B
actionTypeId: .inference
exposeConfig: true
config:
provider: "elastic"
taskType: "chat_completion"
inferenceId: ".gp-llm-v3-chat_completion"
inferenceId: ".openai-gpt-oss-120b-chat_completion"
providerConfig:
model_id: "gp-llm-v3"
model_id: "openai-gpt-oss-120b"
16 changes: 8 additions & 8 deletions config/serverless.security.search_ai_lake.yml
Original file line number Diff line number Diff line change
Expand Up @@ -96,8 +96,8 @@ xpack.fleet.internal.registry.searchAiLakePackageAllowlistEnabled: true

# Elastic Managed LLMs
xpack.actions.preconfigured:
General-Purpose-LLM-v1:
name: General Purpose LLM v1
Anthropic-Claude-Sonnet-3-7:
name: Anthropic Claude Sonnet 3.7
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -106,8 +106,8 @@ xpack.actions.preconfigured:
inferenceId: ".rainbow-sprinkles-elastic"
providerConfig:
model_id: "rainbow-sprinkles"
General-Purpose-LLM-v2:
name: General Purpose LLM v2
Anthropic-Claude-Sonnet-4-5:
name: Anthropic Claude Sonnet 4.5
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -116,13 +116,13 @@ xpack.actions.preconfigured:
inferenceId: ".gp-llm-v2-chat_completion"
providerConfig:
model_id: "gp-llm-v2"
General-Purpose-LLM-v3:
name: General Purpose LLM v3
OpenAI-GPT-OSS-120B:
name: OpenAI GPT-OSS 120B
actionTypeId: .inference
exposeConfig: true
config:
provider: "elastic"
taskType: "chat_completion"
inferenceId: ".gp-llm-v3-chat_completion"
inferenceId: ".openai-gpt-oss-120b-chat_completion"
providerConfig:
model_id: "gp-llm-v3"
model_id: "openai-gpt-oss-120b"
16 changes: 8 additions & 8 deletions config/serverless.workplaceai.yml
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,8 @@ xpack.product_intercept.enabled: false

# Elastic Managed LLMs
xpack.actions.preconfigured:
General-Purpose-LLM-v1:
name: General Purpose LLM v1
Anthropic-Claude-Sonnet-3-7:
name: Anthropic Claude Sonnet 3.7
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -41,8 +41,8 @@ xpack.actions.preconfigured:
inferenceId: ".rainbow-sprinkles-elastic"
providerConfig:
model_id: "rainbow-sprinkles"
General-Purpose-LLM-v2:
name: General Purpose LLM v2
Anthropic-Claude-Sonnet-4-5:
name: Anthropic Claude Sonnet 4.5
actionTypeId: .inference
exposeConfig: true
config:
Expand All @@ -51,13 +51,13 @@ xpack.actions.preconfigured:
inferenceId: ".gp-llm-v2-chat_completion"
providerConfig:
model_id: "gp-llm-v2"
General-Purpose-LLM-v3:
name: General Purpose LLM v3
OpenAI-GPT-OSS-120B:
name: OpenAI GPT-OSS 120B
actionTypeId: .inference
exposeConfig: true
config:
provider: "elastic"
taskType: "chat_completion"
inferenceId: ".gp-llm-v3-chat_completion"
inferenceId: ".openai-gpt-oss-120b-chat_completion"
providerConfig:
model_id: "gp-llm-v3"
model_id: "openai-gpt-oss-120b"
Original file line number Diff line number Diff line change
Expand Up @@ -38,7 +38,7 @@ export const getMessageFromRawResponse = (
}
};

const ELASTIC_LLM_CONNECTOR_IDS = ['Elastic-Managed-LLM', 'General-Purpose-LLM-v1'];
const ELASTIC_LLM_CONNECTOR_IDS = ['Elastic-Managed-LLM', 'Anthropic-Claude-Sonnet-3-7'];

/**
* Returns a default connector if there is only one connector
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,9 +27,9 @@ export const request = <T = unknown>({
};
export const INTERNAL_INFERENCE_CONNECTORS = [
'Elastic-Managed-LLM',
'General-Purpose-LLM-v1',
'General-Purpose-LLM-v2',
'General-Purpose-LLM-v3',
'Anthropic-Claude-Sonnet-3-7',
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Is there a central place we can define this as a constant and import it into the Solutions code, so updates only require a single change rather than multiple edits across the codebase?

'Anthropic-Claude-Sonnet-4-5',
'OpenAI-GPT-OSS-120B',
];
export const INTERNAL_CLOUD_CONNECTORS = ['Elastic-Cloud-SMTP'];

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -19,9 +19,9 @@ import {
const INTERNAL_CLOUD_CONNECTORS = ['Elastic-Cloud-SMTP'];
const INTERNAL_INFERENCE_CONNECTORS = [
'Elastic-Managed-LLM',
'General-Purpose-LLM-v1',
'General-Purpose-LLM-v2',
'General-Purpose-LLM-v3',
'Anthropic-Claude-Sonnet-3-7',
'Anthropic-Claude-Sonnet-4-5',
'OpenAI-GPT-OSS-120B',
];
const INTERNAL_CONNECTORS = [...INTERNAL_CLOUD_CONNECTORS, ...INTERNAL_INFERENCE_CONNECTORS];

Expand Down
Loading