-
Notifications
You must be signed in to change notification settings - Fork 1.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
qwen 2-1.5B model build error #2420
Comments
hardware/gpu H20/tensorrt llm version release 0.14.0 |
Hi @rexmxw02 we've fixed this issue on main branch instead of 0.14 release. Please have a try on main branch. |
convert is ok,but build error: [11/07/2024-02:15:00] [TRT-LLM] [W] padding removal and fMHA are both enabled, max_input_len is not required and will be ignored Invoked with: <tensorrt.tensorrt.INetworkDefinition object at 0x7fdb1d530730>, [<tensorrt.tensorrt.ITensor object at 0x7fdb1d570d30>, <tensorrt.tensorrt.ITensor object at 0x7fd8da553170>], None |
Hi @rexmxw02 , could u please try the latest code base to see if the issue still exists or not? |
System Info
TensorRT-LLM] TensorRT-LLM version: 0.14.0
0.14.0
229it [00:59, 3.88it/s]
Traceback (most recent call last):
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 303, in
main()
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 295, in main
convert_and_save_hf(args)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 251, in convert_and_save_hf
execute(args.workers, [convert_and_save_rank] * world_size, args)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 258, in execute
f(args, rank)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 241, in convert_and_save_rank
qwen = QWenForCausalLM.from_hugging_face(
File "/app/tensorrt_llm/tensorrt_llm/models/qwen/model.py", line 428, in from_hugging_face
loader.generate_tllm_weights(model)
File "/app/tensorrt_llm/tensorrt_llm/models/model_weights_loader.py", line 408, in generate_tllm_weights
self.load(tllm_key,
File "/app/tensorrt_llm/tensorrt_llm/models/model_weights_loader.py", line 296, in load
v = sub_module.postprocess(tllm_key, v, **postprocess_kwargs)
File "/app/tensorrt_llm/tensorrt_llm/layers/linear.py", line 407, in postprocess
weights = weights.to(str_dtype_to_torch(self.dtype))
AttributeError: 'NoneType' object has no attribute 'to'
Who can help?
No response
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
python3 /app/tensorrt_llm/examples/qwen/convert_checkpoint.py --model_dir /Qwen2-1.5B-Instruct --output_dir /qwen --dtype float16 --tp_size 1 --pp_size 1
Expected behavior
python3 /app/tensorrt_llm/examples/qwen/convert_checkpoint.py --model_dir /Qwen2-1.5B-Instruct --output_dir /qwen --dtype float16 --tp_size 1 --pp_size 1
actual behavior
python3 /app/tensorrt_llm/examples/qwen/convert_checkpoint.py --model_dir /Qwen2-1.5B-Instruct --output_dir /qwen --dtype float16 --tp_size 1 --pp_size 1
TensorRT-LLM] TensorRT-LLM version: 0.14.0
0.14.0
229it [00:59, 3.88it/s]
Traceback (most recent call last):
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 303, in
main()
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 295, in main
convert_and_save_hf(args)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 251, in convert_and_save_hf
execute(args.workers, [convert_and_save_rank] * world_size, args)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 258, in execute
f(args, rank)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 241, in convert_and_save_rank
qwen = QWenForCausalLM.from_hugging_face(
File "/app/tensorrt_llm/tensorrt_llm/models/qwen/model.py", line 428, in from_hugging_face
loader.generate_tllm_weights(model)
File "/app/tensorrt_llm/tensorrt_llm/models/model_weights_loader.py", line 408, in generate_tllm_weights
self.load(tllm_key,
File "/app/tensorrt_llm/tensorrt_llm/models/model_weights_loader.py", line 296, in load
v = sub_module.postprocess(tllm_key, v, **postprocess_kwargs)
File "/app/tensorrt_llm/tensorrt_llm/layers/linear.py", line 407, in postprocess
weights = weights.to(str_dtype_to_torch(self.dtype))
AttributeError: 'NoneType' object has no attribute 'to'
additional notes
TensorRT-LLM] TensorRT-LLM version: 0.14.0
0.14.0
229it [00:59, 3.88it/s]
Traceback (most recent call last):
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 303, in
main()
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 295, in main
convert_and_save_hf(args)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 251, in convert_and_save_hf
execute(args.workers, [convert_and_save_rank] * world_size, args)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 258, in execute
f(args, rank)
File "/app/tensorrt_llm/examples/qwen/convert_checkpoint.py", line 241, in convert_and_save_rank
qwen = QWenForCausalLM.from_hugging_face(
File "/app/tensorrt_llm/tensorrt_llm/models/qwen/model.py", line 428, in from_hugging_face
loader.generate_tllm_weights(model)
File "/app/tensorrt_llm/tensorrt_llm/models/model_weights_loader.py", line 408, in generate_tllm_weights
self.load(tllm_key,
File "/app/tensorrt_llm/tensorrt_llm/models/model_weights_loader.py", line 296, in load
v = sub_module.postprocess(tllm_key, v, **postprocess_kwargs)
File "/app/tensorrt_llm/tensorrt_llm/layers/linear.py", line 407, in postprocess
weights = weights.to(str_dtype_to_torch(self.dtype))
AttributeError: 'NoneType' object has no attribute 'to'
The text was updated successfully, but these errors were encountered: