-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What is the right data for SOURCE_LORA
when Convert LoRA to cpp format
#1453
Comments
SOURCE_LORA
when Convert LoRA to cpp format
SOURCE_LORA
when Convert LoRA to cpp format
The nemo converter is only used for nemo ckpt, you cannot use it on huggingface ckpt. |
So what should we do at this step ? I want to benchmark trt-llm for multi lora, I get stuck here. |
This is a miss of the document. You should use # Convert LoRA to cpp format
python examples/hf_lora_convert.py \
-i $SOURCE_LORA \
--storage-type $DTYPE \
-o $CPP_LORA Also, the |
|
#1552 |
I'm benchmarking lora following this: https://github.com/NVIDIA/TensorRT-LLM/blob/main/benchmarks/cpp/README.md#benchmarking-lora
I clone this repo:
https://huggingface.co/hfl/chinese-llama-2-lora-13b
when I run the script:
It reports error.
It looks like the file need to open a tar file, but this model is a folder.
And also, the lora weight folder contains no
model_weights.ckpt
and nomodel_config.yaml
.How can I run this script? Any advice ?
The text was updated successfully, but these errors were encountered: