-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathdocker-compose.yml
More file actions
49 lines (46 loc) · 1.35 KB
/
docker-compose.yml
File metadata and controls
49 lines (46 loc) · 1.35 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
# App + Ollama LLM server. The app uses the Python ollama client (already in the image)
# and connects to the Ollama service via OLLAMA_HOST.
#
# Build and run:
# docker compose build
# docker compose up
#
# Then open the Flet GUI (see below for web vs desktop). Pull a model in Ollama:
# docker compose exec ollama ollama pull llama3.2
services:
ollama:
image: ollama/ollama:latest
container_name: ollama
volumes:
- ollama_data:/root/.ollama
ports:
- "11434:11434"
# Optional: use GPU (uncomment and adjust for your driver)
# deploy:
# resources:
# reservations:
# devices:
# - driver: nvidia
# count: 1
# capabilities: [gpu]
restart: unless-stopped
app:
build: .
container_name: ai-taskvector
environment:
# Point the app at the Ollama service (hostname = service name)
- OLLAMA_HOST=http://ollama:11434
# Run Flet as web server (avoids flet CLI / flet_desktop in container)
- FLET_WEB=1
- FLET_SERVER_PORT=8550
# Optional: default model name
# - OLLAMA_MODEL=llama3.2
depends_on:
- ollama
# For Flet web UI (no display needed): run via Python so we don't need flet CLI / flet_desktop
ports:
- "8550:8550"
command: ["python", "-m", "gui.main"]
restart: "no"
volumes:
ollama_data: