Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

"bot-in-a-box" - model d/l and automatic install into a OpenBLAS or CuBLAS Docker image #270

Merged
merged 6 commits into from
May 25, 2023

Conversation

gjmulder
Copy link
Contributor

@gjmulder gjmulder commented May 23, 2023

This reorganises the Dockerfiles into a ./docker subdir, and provides a Python script for downloading models from 🤗. No models are directly linked to in the source.

There's a ./docker/README.md included which should document everything I've added.

I've renamed and moved the old Dockerfiles as they're still required if you want to externally mount the models as a volume into the Docker image. The new automatic install is for people who just want a "bot-in-a-box" single Docker image with both the latest llama-cpp-python package and a model and don't care about disk space.

@gjmulder gjmulder added build hardware Hardware specific issue labels May 23, 2023
@gjmulder gjmulder changed the title Model d/l and automatic model install into a Docker image Model d/l and automatic install into a Docker image May 23, 2023
@gjmulder gjmulder changed the title Model d/l and automatic install into a Docker image Model d/l and automatic install into a OpenBLAS or CuBLAS Docker image "bot-in-a-box" May 23, 2023
@gjmulder gjmulder changed the title Model d/l and automatic install into a OpenBLAS or CuBLAS Docker image "bot-in-a-box" "bot-in-a-box" - model d/l and automatic install into a OpenBLAS or CuBLAS Docker image May 23, 2023
@abetlen abetlen merged commit de8d9a8 into abetlen:main May 25, 2023
xaptronic pushed a commit to xaptronic/llama-cpp-python that referenced this pull request Jun 13, 2023
* Compute perplexity over prompt

* More accurate perplexity calculation - over all logits in the context window (so 512x more tokens!)

* Output all perplexitiies

* Add timing/ETA
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
build hardware Hardware specific issue
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants