Skip to content
Merged
Show file tree
Hide file tree
Changes from 2 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 5 additions & 3 deletions docs/content/advanced/iac.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,10 +20,12 @@ While the **IaC** can be run from a command-line with prior experience, the step

## Virtual Machine

The Virtual Machine (VM) deployment provisions both the {{< short_app_ref >}} API Server and GUI Client together in an "All-in-One" configuration for experimentation and development. As part of the deployment, one local Large Language Model and one Embedding Model is made available out-of-the-box. There will be an option to deploy on a **GPU**, which will be more expensive then a **CPU** but will perform much better with the pre-deployed Models.
The Virtual Machine (VM) deployment provisions both the {{< short_app_ref >}} API Server and GUI Client together in an "All-in-One" configuration for experimentation and development.

{{% notice style="code" title="Soooo Sloooow..." icon="traffic-light" %}}
If deploying the VM IaC on a **CPU**, we recommend [configuring additional, external models](/client/configuration/model_config) for better performance.
There will be an option to deploy on a **GPU**, which will be more expensive then a **CPU** but will, as part of the deployment, make available one local Large Language Model and one Embedding Model for use out-of-the-box.

{{% notice style="code" title="Models Needed!" icon="traffic-light" %}}
If deploying the VM IaC on a **CPU**, you will need to [configuring a model](/client/configuration/model_config) for functionality.
{{% /notice %}}

### Configure Variables
Expand Down
1 change: 1 addition & 0 deletions opentofu/modules/vm/locals.tf
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ locals {
oci_region = var.region
db_name = var.adb_name
db_password = var.adb_password
install_ollama = var.vm_is_gpu_shape ? true : false
})

vm_compute_shape = var.vm_is_gpu_shape ? var.compute_gpu_shape : var.compute_cpu_shape
Expand Down
20 changes: 13 additions & 7 deletions opentofu/modules/vm/templates/cloudinit-compute.tpl
Original file line number Diff line number Diff line change
Expand Up @@ -40,10 +40,12 @@ write_files:
#!/bin/env bash
mkdir -p /app
chown oracleai:oracleai /app
curl -fsSL https://ollama.com/install.sh | sh
systemctl enable ollama
systemctl daemon-reload
systemctl restart ollama
if ${install_ollama}; then
curl -fsSL https://ollama.com/install.sh | sh
systemctl enable ollama
systemctl daemon-reload
systemctl restart ollama
fi
systemctl stop firewalld.service
firewall-offline-cmd --zone=public --add-port 8501/tcp
firewall-offline-cmd --zone=public --add-port 8000/tcp
Expand Down Expand Up @@ -85,8 +87,10 @@ write_files:
unzip -o /tmp/wallet.zip -d /app/tns_admin

# Install Models
ollama pull llama3.1
ollama pull mxbai-embed-large
if ${install_ollama}; then
ollama pull llama3.1
ollama pull mxbai-embed-large
fi

# Wait for python modules to finish
wait $INSTALL_PID
Expand All @@ -100,7 +104,9 @@ write_files:
export DB_PASSWORD='${db_password}'
export DB_DSN='${db_name}_TP'
export DB_WALLET_PASSWORD='${db_password}'
export ON_PREM_OLLAMA_URL=http://127.0.0.1:11434
if ${install_ollama}; then
export ON_PREM_OLLAMA_URL=http://127.0.0.1:11434
fi
# Clean Cache
find /app -type d -name "__pycache__" -exec rm -rf {} \;
find /app -type d -name ".numba_cache" -exec rm -rf {} \;
Expand Down
2 changes: 1 addition & 1 deletion opentofu/provider.tf
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ terraform {
required_providers {
oci = {
source = "oracle/oci"
version = "~> 7.7" // Last evaluated 1-Jul-2025
version = "~> 7.8" // Last evaluated 8-Jul-2025
}
}
required_version = ">= 1.5"
Expand Down
2 changes: 1 addition & 1 deletion src/server/bootstrap/model_def.py
Original file line number Diff line number Diff line change
Expand Up @@ -159,7 +159,7 @@ def main() -> list[Model]:
"url": os.environ.get("ON_PREM_OLLAMA_URL", default="http://127.0.0.1:11434"),
"api_key": "",
"openai_compat": True,
"max_chunk_size": 512,
"max_chunk_size": 8192,
},
]

Expand Down