Manage your LLM inference server containers.
The backend app uses docker-compose.yml to configure the connection with the host, this included the persistent storage volume, ports, etc
docker compose up
The startup.sh
script automates the management of the environment variables and docker networks that must be configured outside of docker compose.
Environment variables are defined in .env
, .env.default
is a template you can use
cp .env.default .env
# edit JWT_SECRET
vim .env
Note: the backend runs inside a container, because of this it does not have access to the host file system directly to programmatically determine it's relative path.
To remove all containers
# this stops all containers
docker stop $(docker ps -q)
# this deletes all stopped containers
docker container prune
Run the backend and frontend server interactively:
docker compose up
To force rebuilding the Docker images:
docker compose up --build
The local files in ./api
are mounted to /api
within the container for development. You can add breakpoints in the code, it will rebuild and deploy the Django server automatically.
./manage.py runserver 0.0.0.0:8000
For detailed instructions on setting up and running Llama models, including Llama 3.1 70B in TT Studio, refer to this guide