From 9f76b73891d393f4951be5fe4367e8c8c9fab5f4 Mon Sep 17 00:00:00 2001 From: Melissa Date: Wed, 12 Nov 2025 16:29:07 -0700 Subject: [PATCH 1/5] Add preconfigured connector and update name of Elastic Managed LLM preconfig connector --- config/serverless.es.yml | 16 +++++++++++++--- config/serverless.oblt.complete.yml | 16 +++++++++++++--- config/serverless.security.complete.yml | 16 +++++++++++++--- config/serverless.security.search_ai_lake.yml | 16 +++++++++++++--- .../impl/assistant/helpers.ts | 2 +- .../fleet/cypress/tasks/api_calls/connectors.ts | 6 +++++- .../public/management/cypress/tasks/insights.ts | 6 +++++- 7 files changed, 63 insertions(+), 15 deletions(-) diff --git a/config/serverless.es.yml b/config/serverless.es.yml index 73748cf4343e8..f2c5c6cdc059e 100644 --- a/config/serverless.es.yml +++ b/config/serverless.es.yml @@ -114,10 +114,20 @@ xpack.searchQueryRules.enabled: true ## Search Connectors in stack management xpack.contentConnectors.ui.enabled: false -# Elastic Managed LLM +# Elastic Managed LLMs xpack.actions.preconfigured: - Elastic-Managed-LLM: - name: Elastic Managed LLM + General-Purpose-LLM-v2: + name: General Purpose LLM v2 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: "gp-llm-v2" + providerConfig: + model_id: "gp-llm-v2" + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: diff --git a/config/serverless.oblt.complete.yml b/config/serverless.oblt.complete.yml index f1a68b39df09c..e572df527b9c2 100644 --- a/config/serverless.oblt.complete.yml +++ b/config/serverless.oblt.complete.yml @@ -23,10 +23,20 @@ xpack.features.overrides: ### Workflows Management should be moved from Analytics category to the Observability one. workflowsManagement.category: 'observability' -# Elastic Managed LLM +# Elastic Managed LLMs xpack.actions.preconfigured: - Elastic-Managed-LLM: - name: Elastic Managed LLM + General-Purpose-LLM-v2: + name: General Purpose LLM v2 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: "gp-llm-v2" + providerConfig: + model_id: "gp-llm-v2" + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: diff --git a/config/serverless.security.complete.yml b/config/serverless.security.complete.yml index abfeec062cf24..bb41e3005c3ac 100644 --- a/config/serverless.security.complete.yml +++ b/config/serverless.security.complete.yml @@ -7,10 +7,20 @@ xpack.features.overrides: ### Workflows Management should be moved from Analytics category to the Security one. workflowsManagement.category: "security" -# Elastic Managed LLM +# Elastic Managed LLMs xpack.actions.preconfigured: - Elastic-Managed-LLM: - name: Elastic Managed LLM + General-Purpose-LLM-v2: + name: General Purpose LLM v2 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: "gp-llm-v2" + providerConfig: + model_id: "gp-llm-v2" + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: diff --git a/config/serverless.security.search_ai_lake.yml b/config/serverless.security.search_ai_lake.yml index 06a5772324305..96336c48ac809 100644 --- a/config/serverless.security.search_ai_lake.yml +++ b/config/serverless.security.search_ai_lake.yml @@ -94,10 +94,20 @@ xpack.fleet.integrationsHomeOverride: '/app/security/configurations/integrations xpack.fleet.prereleaseEnabledByDefault: true xpack.fleet.internal.registry.searchAiLakePackageAllowlistEnabled: true -# Elastic Managed LLM +# Elastic Managed LLMs xpack.actions.preconfigured: - Elastic-Managed-LLM: - name: Elastic Managed LLM + General-Purpose-LLM-v2: + name: General Purpose LLM v2 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: "gp-llm-v2" + providerConfig: + model_id: "gp-llm-v2" + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: diff --git a/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts b/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts index ff704f0bd0ecf..20dd3c30dd837 100644 --- a/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts +++ b/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts @@ -38,7 +38,7 @@ export const getMessageFromRawResponse = ( } }; -const ELASTIC_LLM_CONNECTOR_ID = 'Elastic-Managed-LLM'; +const ELASTIC_LLM_CONNECTOR_ID = 'General-Purpose-LLM-v1'; /** * Returns a default connector if there is only one connector diff --git a/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts b/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts index 90c76b308e5db..465d36a23f5a4 100644 --- a/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts +++ b/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts @@ -25,7 +25,11 @@ export const request = ({ ...options, }); }; -export const INTERNAL_INFERENCE_CONNECTORS = ['Elastic-Managed-LLM']; +export const INTERNAL_INFERENCE_CONNECTORS = [ + 'Elastic-Managed-LLM', + 'General-Purpose-LLM-v1', + 'General-Purpose-LLM-v2', +]; export const INTERNAL_CLOUD_CONNECTORS = ['Elastic-Cloud-SMTP']; export const getConnectors = () => diff --git a/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts b/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts index 55d1fc4bc2ef6..c652e08db1526 100644 --- a/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts +++ b/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts @@ -17,7 +17,11 @@ import { } from '../../../../common/endpoint/constants'; const INTERNAL_CLOUD_CONNECTORS = ['Elastic-Cloud-SMTP']; -const INTERNAL_INFERENCE_CONNECTORS = ['Elastic-Managed-LLM']; +const INTERNAL_INFERENCE_CONNECTORS = [ + 'Elastic-Managed-LLM', + 'General-Purpose-LLM-v1', + 'General-Purpose-LLM-v2', +]; const INTERNAL_CONNECTORS = [...INTERNAL_CLOUD_CONNECTORS, ...INTERNAL_INFERENCE_CONNECTORS]; export const createBedrockAIConnector = (connectorName?: string) => From 37451e296fc5ca55ac4e841ed23b8177148e56fa Mon Sep 17 00:00:00 2001 From: Melissa Date: Thu, 13 Nov 2025 11:22:51 -0700 Subject: [PATCH 2/5] add llm-v3 and update inference endpoints --- config/serverless.es.yml | 12 +++++++++++- config/serverless.oblt.complete.yml | 12 +++++++++++- config/serverless.security.complete.yml | 12 +++++++++++- config/serverless.security.search_ai_lake.yml | 12 +++++++++++- .../fleet/cypress/tasks/api_calls/connectors.ts | 1 + .../public/management/cypress/tasks/insights.ts | 1 + 6 files changed, 46 insertions(+), 4 deletions(-) diff --git a/config/serverless.es.yml b/config/serverless.es.yml index f2c5c6cdc059e..ee7dc9725e9b1 100644 --- a/config/serverless.es.yml +++ b/config/serverless.es.yml @@ -116,6 +116,16 @@ xpack.contentConnectors.ui.enabled: false # Elastic Managed LLMs xpack.actions.preconfigured: + General-Purpose-LLM-v3: + name: General Purpose LLM v3 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v3-chat_completion" + providerConfig: + model_id: "gp-llm-v3" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -123,7 +133,7 @@ xpack.actions.preconfigured: config: provider: "elastic" taskType: "chat_completion" - inferenceId: "gp-llm-v2" + inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" General-Purpose-LLM-v1: diff --git a/config/serverless.oblt.complete.yml b/config/serverless.oblt.complete.yml index e572df527b9c2..d0089d82088cc 100644 --- a/config/serverless.oblt.complete.yml +++ b/config/serverless.oblt.complete.yml @@ -25,6 +25,16 @@ xpack.features.overrides: # Elastic Managed LLMs xpack.actions.preconfigured: + General-Purpose-LLM-v3: + name: General Purpose LLM v3 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v3-chat_completion" + providerConfig: + model_id: "gp-llm-v3" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -32,7 +42,7 @@ xpack.actions.preconfigured: config: provider: "elastic" taskType: "chat_completion" - inferenceId: "gp-llm-v2" + inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" General-Purpose-LLM-v1: diff --git a/config/serverless.security.complete.yml b/config/serverless.security.complete.yml index bb41e3005c3ac..76fb8e87c5248 100644 --- a/config/serverless.security.complete.yml +++ b/config/serverless.security.complete.yml @@ -9,6 +9,16 @@ xpack.features.overrides: # Elastic Managed LLMs xpack.actions.preconfigured: + General-Purpose-LLM-v3: + name: General Purpose LLM v3 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v3-chat_completion" + providerConfig: + model_id: "gp-llm-v3" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -16,7 +26,7 @@ xpack.actions.preconfigured: config: provider: "elastic" taskType: "chat_completion" - inferenceId: "gp-llm-v2" + inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" General-Purpose-LLM-v1: diff --git a/config/serverless.security.search_ai_lake.yml b/config/serverless.security.search_ai_lake.yml index 96336c48ac809..df0526beeca40 100644 --- a/config/serverless.security.search_ai_lake.yml +++ b/config/serverless.security.search_ai_lake.yml @@ -96,6 +96,16 @@ xpack.fleet.internal.registry.searchAiLakePackageAllowlistEnabled: true # Elastic Managed LLMs xpack.actions.preconfigured: + General-Purpose-LLM-v3: + name: General Purpose LLM v3 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v3-chat_completion" + providerConfig: + model_id: "gp-llm-v3" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -103,7 +113,7 @@ xpack.actions.preconfigured: config: provider: "elastic" taskType: "chat_completion" - inferenceId: "gp-llm-v2" + inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" General-Purpose-LLM-v1: diff --git a/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts b/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts index 465d36a23f5a4..2f21ec5ed4df8 100644 --- a/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts +++ b/x-pack/platform/plugins/shared/fleet/cypress/tasks/api_calls/connectors.ts @@ -29,6 +29,7 @@ export const INTERNAL_INFERENCE_CONNECTORS = [ 'Elastic-Managed-LLM', 'General-Purpose-LLM-v1', 'General-Purpose-LLM-v2', + 'General-Purpose-LLM-v3', ]; export const INTERNAL_CLOUD_CONNECTORS = ['Elastic-Cloud-SMTP']; diff --git a/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts b/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts index c652e08db1526..4447a21b2811a 100644 --- a/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts +++ b/x-pack/solutions/security/plugins/security_solution/public/management/cypress/tasks/insights.ts @@ -21,6 +21,7 @@ const INTERNAL_INFERENCE_CONNECTORS = [ 'Elastic-Managed-LLM', 'General-Purpose-LLM-v1', 'General-Purpose-LLM-v2', + 'General-Purpose-LLM-v3', ]; const INTERNAL_CONNECTORS = [...INTERNAL_CLOUD_CONNECTORS, ...INTERNAL_INFERENCE_CONNECTORS]; From 71e8005876ff2dd1c8742769dc619067cdfae3b9 Mon Sep 17 00:00:00 2001 From: Melissa Date: Tue, 18 Nov 2025 12:36:46 -0700 Subject: [PATCH 3/5] check legacy name as well as new name --- .../shared/kbn-elastic-assistant/impl/assistant/helpers.ts | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts b/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts index 20dd3c30dd837..8ada4916d3e09 100644 --- a/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts +++ b/x-pack/platform/packages/shared/kbn-elastic-assistant/impl/assistant/helpers.ts @@ -38,7 +38,7 @@ export const getMessageFromRawResponse = ( } }; -const ELASTIC_LLM_CONNECTOR_ID = 'General-Purpose-LLM-v1'; +const ELASTIC_LLM_CONNECTOR_IDS = ['Elastic-Managed-LLM', 'General-Purpose-LLM-v1']; /** * Returns a default connector if there is only one connector @@ -67,7 +67,7 @@ export const getDefaultConnector = ( // In case the default connector is not set or is invalid, return the prioritized connector const prioritizedConnectors = [...validConnectors].sort((a, b) => { const priority = (connector: (typeof validConnectors)[number]) => { - if (connector.id === ELASTIC_LLM_CONNECTOR_ID) return 0; + if (ELASTIC_LLM_CONNECTOR_IDS.includes(connector.id)) return 0; if ( connector.apiProvider === OpenAiProviderType.OpenAi || connector.apiProvider === OpenAiProviderType.AzureAi From 96c422cd3cdae43e4351ac41c7d40d5f54384714 Mon Sep 17 00:00:00 2001 From: Melissa Date: Tue, 25 Nov 2025 10:00:49 -0700 Subject: [PATCH 4/5] update workplaceai.yml file --- config/serverless.workplaceai.yml | 26 +++++++++++++++++++++++--- 1 file changed, 23 insertions(+), 3 deletions(-) diff --git a/config/serverless.workplaceai.yml b/config/serverless.workplaceai.yml index ec731b1bd3ee1..9be1fe283e44a 100644 --- a/config/serverless.workplaceai.yml +++ b/config/serverless.workplaceai.yml @@ -29,10 +29,30 @@ xpack.contentConnectors.enabled: false ## Disable Kibana Product Intercept xpack.product_intercept.enabled: false -# Elastic Managed LLM +# Elastic Managed LLMs xpack.actions.preconfigured: - Elastic-Managed-LLM: - name: Elastic Managed LLM + General-Purpose-LLM-v3: + name: General Purpose LLM v3 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v3-chat_completion" + providerConfig: + model_id: "gp-llm-v3" + General-Purpose-LLM-v2: + name: General Purpose LLM v2 + actionTypeId: .inference + exposeConfig: true + config: + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v2-chat_completion" + providerConfig: + model_id: "gp-llm-v2" + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: From 0146ed43211bf5afb9b1700f77936ebbb906fe83 Mon Sep 17 00:00:00 2001 From: Melissa Date: Tue, 25 Nov 2025 13:53:29 -0700 Subject: [PATCH 5/5] update order of preconfig connectors --- config/serverless.es.yml | 16 +++++++-------- config/serverless.oblt.complete.yml | 20 +++++++++---------- config/serverless.security.complete.yml | 16 +++++++-------- config/serverless.security.search_ai_lake.yml | 16 +++++++-------- config/serverless.workplaceai.yml | 16 +++++++-------- 5 files changed, 42 insertions(+), 42 deletions(-) diff --git a/config/serverless.es.yml b/config/serverless.es.yml index ee7dc9725e9b1..0107fbfcf2cf1 100644 --- a/config/serverless.es.yml +++ b/config/serverless.es.yml @@ -116,16 +116,16 @@ xpack.contentConnectors.ui.enabled: false # Elastic Managed LLMs xpack.actions.preconfigured: - General-Purpose-LLM-v3: - name: General Purpose LLM v3 + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".gp-llm-v3-chat_completion" + inferenceId: ".rainbow-sprinkles-elastic" providerConfig: - model_id: "gp-llm-v3" + model_id: "rainbow-sprinkles" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -136,13 +136,13 @@ xpack.actions.preconfigured: inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" - General-Purpose-LLM-v1: - name: General Purpose LLM v1 + General-Purpose-LLM-v3: + name: General Purpose LLM v3 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".rainbow-sprinkles-elastic" + inferenceId: ".gp-llm-v3-chat_completion" providerConfig: - model_id: "rainbow-sprinkles" + model_id: "gp-llm-v3" diff --git a/config/serverless.oblt.complete.yml b/config/serverless.oblt.complete.yml index d0089d82088cc..348b2f04933dd 100644 --- a/config/serverless.oblt.complete.yml +++ b/config/serverless.oblt.complete.yml @@ -25,16 +25,16 @@ xpack.features.overrides: # Elastic Managed LLMs xpack.actions.preconfigured: - General-Purpose-LLM-v3: - name: General Purpose LLM v3 + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".gp-llm-v3-chat_completion" + inferenceId: ".rainbow-sprinkles-elastic" providerConfig: - model_id: "gp-llm-v3" + model_id: "rainbow-sprinkles" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -45,13 +45,13 @@ xpack.actions.preconfigured: inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" - General-Purpose-LLM-v1: - name: General Purpose LLM v1 + General-Purpose-LLM-v3: + name: General Purpose LLM v3 actionTypeId: .inference exposeConfig: true config: - provider: 'elastic' - taskType: 'chat_completion' - inferenceId: '.rainbow-sprinkles-elastic' + provider: "elastic" + taskType: "chat_completion" + inferenceId: ".gp-llm-v3-chat_completion" providerConfig: - model_id: 'rainbow-sprinkles' + model_id: "gp-llm-v3" diff --git a/config/serverless.security.complete.yml b/config/serverless.security.complete.yml index 76fb8e87c5248..2c98bf7866e1b 100644 --- a/config/serverless.security.complete.yml +++ b/config/serverless.security.complete.yml @@ -9,16 +9,16 @@ xpack.features.overrides: # Elastic Managed LLMs xpack.actions.preconfigured: - General-Purpose-LLM-v3: - name: General Purpose LLM v3 + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".gp-llm-v3-chat_completion" + inferenceId: ".rainbow-sprinkles-elastic" providerConfig: - model_id: "gp-llm-v3" + model_id: "rainbow-sprinkles" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -29,13 +29,13 @@ xpack.actions.preconfigured: inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" - General-Purpose-LLM-v1: - name: General Purpose LLM v1 + General-Purpose-LLM-v3: + name: General Purpose LLM v3 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".rainbow-sprinkles-elastic" + inferenceId: ".gp-llm-v3-chat_completion" providerConfig: - model_id: "rainbow-sprinkles" + model_id: "gp-llm-v3" diff --git a/config/serverless.security.search_ai_lake.yml b/config/serverless.security.search_ai_lake.yml index df0526beeca40..6cd7651fa9b76 100644 --- a/config/serverless.security.search_ai_lake.yml +++ b/config/serverless.security.search_ai_lake.yml @@ -96,16 +96,16 @@ xpack.fleet.internal.registry.searchAiLakePackageAllowlistEnabled: true # Elastic Managed LLMs xpack.actions.preconfigured: - General-Purpose-LLM-v3: - name: General Purpose LLM v3 + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".gp-llm-v3-chat_completion" + inferenceId: ".rainbow-sprinkles-elastic" providerConfig: - model_id: "gp-llm-v3" + model_id: "rainbow-sprinkles" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -116,13 +116,13 @@ xpack.actions.preconfigured: inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" - General-Purpose-LLM-v1: - name: General Purpose LLM v1 + General-Purpose-LLM-v3: + name: General Purpose LLM v3 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".rainbow-sprinkles-elastic" + inferenceId: ".gp-llm-v3-chat_completion" providerConfig: - model_id: "rainbow-sprinkles" + model_id: "gp-llm-v3" diff --git a/config/serverless.workplaceai.yml b/config/serverless.workplaceai.yml index 9be1fe283e44a..04d770fca66f9 100644 --- a/config/serverless.workplaceai.yml +++ b/config/serverless.workplaceai.yml @@ -31,16 +31,16 @@ xpack.product_intercept.enabled: false # Elastic Managed LLMs xpack.actions.preconfigured: - General-Purpose-LLM-v3: - name: General Purpose LLM v3 + General-Purpose-LLM-v1: + name: General Purpose LLM v1 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".gp-llm-v3-chat_completion" + inferenceId: ".rainbow-sprinkles-elastic" providerConfig: - model_id: "gp-llm-v3" + model_id: "rainbow-sprinkles" General-Purpose-LLM-v2: name: General Purpose LLM v2 actionTypeId: .inference @@ -51,13 +51,13 @@ xpack.actions.preconfigured: inferenceId: ".gp-llm-v2-chat_completion" providerConfig: model_id: "gp-llm-v2" - General-Purpose-LLM-v1: - name: General Purpose LLM v1 + General-Purpose-LLM-v3: + name: General Purpose LLM v3 actionTypeId: .inference exposeConfig: true config: provider: "elastic" taskType: "chat_completion" - inferenceId: ".rainbow-sprinkles-elastic" + inferenceId: ".gp-llm-v3-chat_completion" providerConfig: - model_id: "rainbow-sprinkles" + model_id: "gp-llm-v3"