Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Peter/cogvlm #175

Merged
merged 12 commits into from
Dec 7, 2023
Merged

Peter/cogvlm #175

merged 12 commits into from
Dec 7, 2023

Conversation

probicheaux
Copy link
Collaborator

@probicheaux probicheaux commented Dec 7, 2023

Description

Add in CogVLM to roboflow inference server (with chat history!)

See examples/cogvlm/cog_client.py for an example script.
Docker image available at: roboflow/roboflow-inference-server-cog

I ran the image with the command:

docker run --gpus=all --rm -it -v cog:/cache --net=host --env NVIDIA_DISABLE_REQUIRE=1 roboflow/roboflow-inference-server-cog

Dependencies added:

transformers<=4.35.2
sentencepiece<=0.1.99
einops<=0.7.0
xformers<=0.0.22
accelerate<=0.25.0
bitsandbytes<=0.41.2.post2

Type of change

Please delete options that are not relevant.

  • Bug fix (non-breaking change which fixes an issue)
  • New feature (non-breaking change which adds functionality)
  • This change requires a documentation update

How has this change been tested, please provide a testcase or example of how you tested the change?

Ran on a T4.

Any specific deployment considerations

Note the env variables COG_LOAD_8BIT and COG_LOAD_4BIT. COG_LOAD_4BIT must be set to true to run on a T4 (this is the current default).

Docs

  • Docs updated? What were the changes:

Copy link
Contributor

@paulguerrie paulguerrie left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looking great!

inference/models/cogvlm/cog.py Outdated Show resolved Hide resolved
@paulguerrie paulguerrie self-requested a review December 7, 2023 20:37
@probicheaux probicheaux merged commit 15daa63 into main Dec 7, 2023
5 checks passed
@probicheaux probicheaux deleted the peter/cogvlm branch December 7, 2023 21:30
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants