You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
AssertionError: Attempted to load weight (torch.Size([3584])) into parameter (torch.Size([3584, 7168]))
I lookup to the code vllm/model_executor/models/eagle.py:L139 which is shown as below:
I think you only consider the name varieble startswith 'fc.' can only be 'fc.weight', but the fc layer of eagle-qwen2 has bias attribute, which means the name varieble can be 'fc.bias'.
Moreover, the qkv_proj layer of EAGLE-Qwen2-7B-Instruct also have bias.
I hope you can fix this in the upcoming upgrade!
The text was updated successfully, but these errors were encountered:
I can successfully deploy llama3-8b-instruct with EAGLE. But there is a problem when deploying qwen2-7b-instruct with EAGLE.
I have converted the EAGLE-Qwen2-7B-Instruct model according tovllm/model_executor/models/eagle.py:L126.
I encountered another error below:
AssertionError: Attempted to load weight (torch.Size([3584])) into parameter (torch.Size([3584, 7168]))
I lookup to the code vllm/model_executor/models/eagle.py:L139 which is shown as below:
I think you only consider the name varieble startswith 'fc.' can only be 'fc.weight', but the fc layer of eagle-qwen2 has bias attribute, which means the name varieble can be 'fc.bias'.
Moreover, the qkv_proj layer of EAGLE-Qwen2-7B-Instruct also have bias.
I hope you can fix this in the upcoming upgrade!
The text was updated successfully, but these errors were encountered: