Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] Configuration Document Optimization Suggestions - ChatQnA Setup and Deployment #1251

Open
zhangyuting opened this issue Dec 14, 2024 · 0 comments
Labels
feature New feature or request

Comments

@zhangyuting
Copy link

Priority

P3-Medium

OS type

Ubuntu

Hardware type

GPU-Nvidia

Running nodes

Single Node

Description

Due to the limited performance of my local Intel processor, I tried to deploy ChatQnA using a 3090 graphics card. According to the documentation at https://github.com/opea-project/GenAIExamples/blob/main/ChatQnA/docker_compose/nvidia/gpu/README.md, everything went smoothly.

However, after completing the deployment, I found that the host's large language model was still using the CPU instead of the GPU. After some troubleshooting, I discovered that the CUDA version on Ubuntu was incompatible with the torch version inside TGI Docker, which caused the GPU not to load. I saw a great table in the documentation listing which Docker image versions should be used for different GPU series; this is very helpful. I also suggest providing recommendations for which CUDA versions should be installed on Ubuntu hosts so that users can complete their setup more smoothly and quickly transition to developing upper-level applications.
image
image

Overall, OPEA is fantastic.

@zhangyuting zhangyuting added the feature New feature or request label Dec 14, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant