Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

stuck when infer wIth trained lora model #30

Open
moon-fall opened this issue Oct 5, 2023 · 0 comments
Open

stuck when infer wIth trained lora model #30

moon-fall opened this issue Oct 5, 2023 · 0 comments

Comments

@moon-fall
Copy link

infer using 2 A100 80g
run the inference,py script

the log shows below

CUDA_SETUP: WARNING! libcudart.so not found in any environmental path. Searching /usr/local/cuda/lib64...
CUDA SETUP: CUDA runtime path found: /usr/local/cuda/lib64/libcudart.so
CUDA SETUP: Highest compute capability among GPUs detected: 8.0
CUDA SETUP: Detected CUDA version 113
CUDA SETUP: Loading binary /home/notebook/data/group/cubelm/lf_cubelm/lib/python3.9/site-packages/bitsandbytes/libbitsandbytes_cuda113.so...
[2023-10-05 10:09:55,419] [INFO] [real_accelerator.py:110:get_accelerator] Setting ds_accelerator to cuda (auto detect)
Loading checkpoint shards: 100%|████████████████| 15/15 [04:43<00:00, 18.93s/it]
Processing batch 1 of 1...

but using original llama-2-70b model without lora to infer is OK

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant