git-subtree-dir: apps/core git-subtree-mainline: a8963e11e7a5a0059acbc849ce768e1eee80df61 git-subtree-split: ea2a7395f4023f5b9fab03e6273db3b64a1181d5
88 lines
2.4 KiB
YAML
88 lines
2.4 KiB
YAML
name: Endpoint (Ollama)
|
|
|
|
env:
|
|
OLLAMA_BASE_URL: "http://localhost:11434"
|
|
COMPOSIO_API_KEY: ${{ secrets.COMPOSIO_API_KEY }}
|
|
|
|
on:
|
|
push:
|
|
branches: [ main ]
|
|
pull_request:
|
|
branches: [ main ]
|
|
|
|
jobs:
|
|
test:
|
|
runs-on: ubuntu-latest
|
|
timeout-minutes: 15
|
|
steps:
|
|
- name: Checkout Code
|
|
uses: actions/checkout@v4
|
|
|
|
- name: Install Ollama
|
|
run: |
|
|
set -e
|
|
set -x
|
|
curl -vfsSL https://ollama.com/install.sh -o install.sh
|
|
chmod +x install.sh
|
|
bash -x install.sh
|
|
if ! command -v ollama; then
|
|
echo "Ollama binary not found in PATH after installation."
|
|
exit 1
|
|
fi
|
|
echo "Ollama installed successfully."
|
|
|
|
- name: Start Ollama Server
|
|
run: |
|
|
set -e
|
|
set -x
|
|
ollama serve >ollama_server.log 2>&1 &
|
|
sleep 15
|
|
if ! curl -v http://localhost:11434; then
|
|
echo "Server logs (if available):"
|
|
[ -f ollama_server.log ] && cat ollama_server.log || echo "No logs found."
|
|
exit 1
|
|
fi
|
|
echo "Ollama server started successfully."
|
|
|
|
- name: Pull Models
|
|
run: |
|
|
set -e
|
|
set -x
|
|
for attempt in {1..3}; do
|
|
ollama pull thewindmom/hermes-3-llama-3.1-8b && break || sleep 5
|
|
done
|
|
for attempt in {1..3}; do
|
|
ollama pull mxbai-embed-large && break || sleep 5
|
|
done
|
|
|
|
- name: Debug Logs on Failure
|
|
if: failure()
|
|
run: |
|
|
echo "Debugging logs on failure:"
|
|
[ -f ollama_server.log ] && cat ollama_server.log || echo "No server logs available."
|
|
|
|
- name: Setup Python, Poetry, and Dependencies
|
|
uses: packetcoders/action-setup-cache-python-poetry@main
|
|
with:
|
|
python-version: "3.12"
|
|
poetry-version: "1.8.2"
|
|
install-args: "-E dev"
|
|
|
|
- name: Test LLM Endpoint
|
|
run: |
|
|
set -e
|
|
set -x
|
|
poetry run pytest -s -vv tests/test_model_letta_perfomance.py::test_llm_endpoint_ollama
|
|
|
|
- name: Test Embedding Endpoint
|
|
run: |
|
|
set -e
|
|
set -x
|
|
poetry run pytest -s -vv tests/test_model_letta_perfomance.py::test_embedding_endpoint_ollama
|
|
|
|
- name: Test Provider
|
|
run: |
|
|
set -e
|
|
set -x
|
|
poetry run pytest -s -vv tests/test_providers.py::test_ollama
|