This is a Stable Diffusion starter template from Banana.dev that allows on-demand serverless GPU inference.
It uses a custom safetensors model from Hugging Face.
It supports prompting with and without seeds.
You can fork this repository and deploy it on Banana as is, or customize it based on your own needs.
- Fork this repository to your own Github account.
- Connect your Github account on Banana.
- Create a new model on Banana from the forked Github repository.
- Wait for the model to build after creating it.
- Make an API request to it using one of the provided snippets in your Banana dashboard.
For more info, check out the Banana.dev docs.
Build the model as a Docker image. You can change the banana-sd-hf-safetensors
part to anything.
docker build -t banana-sd-hf-safetensors .
Run the Potassium server
docker run --publish 8000:8000 -it banana-sd-hf-safetensors
In another terminal, run inference after the above is built and running.
curl -X POST -H 'Content-Type: application/json' -d '{"prompt": "colorful, 1girl, high angle, software developer protagonist, hacker, ((computer screen)), matrix, ((keyboard)), glow, light particles, wallpaper, chromatic aberration"}' http://localhost:8000
You could also install and run it without Docker.
Just make sure that the pip dependencies in the Docker file (and torch) are installed in your Python virtual environment.
Run the Potassium app in one terminal window.
python3 app.py
Call the model in another terminal window with the Potassium app still running from the previous step.
curl -X POST -H 'Content-Type: application/json' -d '{"prompt": "colorful, 1girl, high angle, software developer protagonist, hacker, ((computer screen)), matrix, ((keyboard)), glow, light particles, wallpaper, chromatic aberration"}' http://localhost:8000