diff --git a/.github/workflows/functionality-helm-chart.yml b/.github/workflows/functionality-helm-chart.yml index 356e88d3e..42bb9541a 100644 --- a/.github/workflows/functionality-helm-chart.yml +++ b/.github/workflows/functionality-helm-chart.yml @@ -66,7 +66,7 @@ jobs: - name: Validate the installation and send query to the stack run: | bash .github/port-forward.sh curl-05-secure-vllm - timeout-minutes: 2 + timeout-minutes: 3 - name: Archive functionality results uses: actions/upload-artifact@v4 if: always() @@ -112,7 +112,7 @@ jobs: - name: Validate the installation and send query to the stack run: | bash .github/port-forward.sh curl-02-two-pods - timeout-minutes: 2 + timeout-minutes: 3 - name: Archive functionality results uses: actions/upload-artifact@v4 if: always() @@ -156,7 +156,7 @@ jobs: - name: Validate the installation and send query to the stack run: | bash .github/port-forward.sh curl-04-multiple-models - timeout-minutes: 3 + timeout-minutes: 5 - name: Archive functionality results uses: actions/upload-artifact@v4 if: always() diff --git a/.github/workflows/router-e2e-test.yml b/.github/workflows/router-e2e-test.yml index b689223ad..85f7c4c47 100644 --- a/.github/workflows/router-e2e-test.yml +++ b/.github/workflows/router-e2e-test.yml @@ -109,13 +109,10 @@ jobs: - name: Check out repository code uses: actions/checkout@v4 - - name: Setup Python - uses: actions/setup-python@v5 - with: - python-version: "3.12" - - name: Install Python dependencies run: | + source /usr/local/bin/conda-init + conda activate llmstack python -m pip install --upgrade pip pip install -r benchmarks/multi-round-qa/requirements.txt pip install -e . @@ -142,6 +139,8 @@ jobs: - name: Run all k8s discovery routing tests run: | + source /usr/local/bin/conda-init + conda activate llmstack echo "๐Ÿงช Running all k8s discovery routing tests" ./tests/e2e/run-k8s-routing-test.sh all \ --model "facebook/opt-125m" \ @@ -173,23 +172,24 @@ jobs: - name: Check out repository code uses: actions/checkout@v4 - - name: Setup Python - uses: actions/setup-python@v5 - with: - python-version: "3.12" - - name: Install Python dependencies run: | + source /usr/local/bin/conda-init + conda activate llmstack python -m pip install --upgrade pip pip install -e . - name: Install vLLM and lmcache run: | + source /usr/local/bin/conda-init + conda activate llmstack pip install vllm pip install lmcache - name: Start 2 vLLM serve backends run: | + source /usr/local/bin/conda-init + conda activate llmstack echo "๐Ÿš€ Starting vLLM serve backend" mkdir -p "$LOG_DIR" CUDA_VISIBLE_DEVICES=0 vllm serve facebook/opt-125m --port 8001 --gpu-memory-utilization 0.7 --chat-template .github/template-chatml.jinja > "$LOG_DIR/backend1.log" 2>&1 & @@ -202,9 +202,9 @@ jobs: ./tests/e2e/wait-for-backends.sh 180 "http://localhost:8001" "http://localhost:8002" - name: Run All Static Discovery Routing Tests - env: - PYTHONPATH: ${{ github.workspace }}/src run: | + source /usr/local/bin/conda-init + conda activate llmstack echo "๐Ÿงช Running all static discovery routing tests sequentially" chmod +x tests/e2e/run-static-discovery-routing-test.sh ./tests/e2e/run-static-discovery-routing-test.sh all \