Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
25 changes: 22 additions & 3 deletions .github/workflows/documentation.yml
Original file line number Diff line number Diff line change
Expand Up @@ -78,11 +78,30 @@ jobs:
--source docs \
--baseURL "${{ steps.pages.outputs.base_url }}/"

- name: Package Helm chart
- name: Package Tags Helm chart
run: |
TAGS=$(git tag --sort=creatordate)
for TAG in $TAGS; do
echo "Processing tag: $TAG"

# Create a temporary worktree to avoid changing the current working directory
WORKDIR=$(mktemp -d)
git worktree add "$WORKDIR" "$TAG"

# Package Helm chart for this tag
helm package "$WORKDIR/helm" -d docs/public/helm --debug

# Clean up
git worktree remove "$WORKDIR"
done

- name: Package Main Helm chart
run: |
mkdir -p docs/public/helm
helm package helm -d docs/public/helm
helm repo index docs/public/helm --url "${{ steps.pages.outputs.base_url }}/helm"
helm package helm -d docs/public/helm --debug

- name: Index Helm charts
run: helm repo index docs/public/helm --url "${{ steps.pages.outputs.base_url }}/helm"

- name: Upload artifact
uses: actions/upload-pages-artifact@v3
Expand Down
8 changes: 5 additions & 3 deletions docs/content/advanced/iac.md
Original file line number Diff line number Diff line change
Expand Up @@ -20,10 +20,12 @@ While the **IaC** can be run from a command-line with prior experience, the step

## Virtual Machine

The Virtual Machine (VM) deployment provisions both the {{< short_app_ref >}} API Server and GUI Client together in an "All-in-One" configuration for experimentation and development. As part of the deployment, one local Large Language Model and one Embedding Model is made available out-of-the-box. There will be an option to deploy on a **GPU**, which will be more expensive then a **CPU** but will perform much better with the pre-deployed Models.
The Virtual Machine (VM) deployment provisions both the {{< short_app_ref >}} API Server and GUI Client together in an "All-in-One" configuration for experimentation and development.

{{% notice style="code" title="Soooo Sloooow..." icon="traffic-light" %}}
If deploying the VM IaC on a **CPU**, we recommend [configuring additional, external models](/client/configuration/model_config) for better performance.
There will be an option to deploy on a **GPU**, which will be more expensive then a **CPU** but will, as part of the deployment, make available one local Large Language Model and one Embedding Model for use out-of-the-box.

{{% notice style="code" title="Models Needed!" icon="traffic-light" %}}
If deploying the VM IaC on a **CPU**, you will need to [configure a model](/client/configuration/model_config) for functionality.
{{% /notice %}}

### Configure Variables
Expand Down
1 change: 1 addition & 0 deletions opentofu/modules/vm/locals.tf
Original file line number Diff line number Diff line change
Expand Up @@ -9,6 +9,7 @@ locals {
oci_region = var.region
db_name = var.adb_name
db_password = var.adb_password
install_ollama = var.vm_is_gpu_shape ? true : false
})

vm_compute_shape = var.vm_is_gpu_shape ? var.compute_gpu_shape : var.compute_cpu_shape
Expand Down
20 changes: 13 additions & 7 deletions opentofu/modules/vm/templates/cloudinit-compute.tpl
Original file line number Diff line number Diff line change
Expand Up @@ -40,10 +40,12 @@ write_files:
#!/bin/env bash
mkdir -p /app
chown oracleai:oracleai /app
curl -fsSL https://ollama.com/install.sh | sh
systemctl enable ollama
systemctl daemon-reload
systemctl restart ollama
if ${install_ollama}; then
curl -fsSL https://ollama.com/install.sh | sh
systemctl enable ollama
systemctl daemon-reload
systemctl restart ollama
fi
systemctl stop firewalld.service
firewall-offline-cmd --zone=public --add-port 8501/tcp
firewall-offline-cmd --zone=public --add-port 8000/tcp
Expand Down Expand Up @@ -85,8 +87,10 @@ write_files:
unzip -o /tmp/wallet.zip -d /app/tns_admin

# Install Models
ollama pull llama3.1
ollama pull mxbai-embed-large
if ${install_ollama}; then
ollama pull llama3.1
ollama pull mxbai-embed-large
fi

# Wait for python modules to finish
wait $INSTALL_PID
Expand All @@ -100,7 +104,9 @@ write_files:
export DB_PASSWORD='${db_password}'
export DB_DSN='${db_name}_TP'
export DB_WALLET_PASSWORD='${db_password}'
export ON_PREM_OLLAMA_URL=http://127.0.0.1:11434
if ${install_ollama}; then
export ON_PREM_OLLAMA_URL=http://127.0.0.1:11434
fi
# Clean Cache
find /app -type d -name "__pycache__" -exec rm -rf {} \;
find /app -type d -name ".numba_cache" -exec rm -rf {} \;
Expand Down
2 changes: 1 addition & 1 deletion opentofu/provider.tf
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ terraform {
required_providers {
oci = {
source = "oracle/oci"
version = "~> 7.7" // Last evaluated 1-Jul-2025
version = "~> 7.8" // Last evaluated 8-Jul-2025
}
}
required_version = ">= 1.5"
Expand Down
2 changes: 1 addition & 1 deletion src/server/bootstrap/model_def.py
Original file line number Diff line number Diff line change
Expand Up @@ -159,7 +159,7 @@ def main() -> list[Model]:
"url": os.environ.get("ON_PREM_OLLAMA_URL", default="http://127.0.0.1:11434"),
"api_key": "",
"openai_compat": True,
"max_chunk_size": 512,
"max_chunk_size": 8192,
},
]

Expand Down