Basic setup to run self-hosted Ollama and use LLMs locally (Works with Coolify)
- Ubuntu 22.04
- Nvidia GPU
- Docker
- Run setup.bash (installs Nvidia drivers, Nvidia CUDA toolkit, and Nvidia Container Toolkit)
- If using Coolify, paste the docker-compose.yaml content in the Docker Compose section of your project. Otherwise, just run
docker compose up -d
inside the project's folder on your machine. - Go to localhost:3000 or the respective Coolify link for your service.
- Model Library: https://ollama.com/library
- Run
docker exec <ollama_container_id> -it ollama pull <modelname>