Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TensorRT-LLM插件设置顺序错误导致engine运行时报错(TensorRT Hackathon 2023) #95

Open
Eddie-Wang1120 opened this issue Sep 20, 2023 · 0 comments

Comments

@Eddie-Wang1120
Copy link

Eddie-Wang1120 commented Sep 20, 2023

Environment
TensorRT 9.0.0.2 EA
Container used: registry.cn-hangzhou.aliyuncs.com/trt-hackathon/trt-hackathon:final_v1
NVIDIA-SMI 525.105.17 Driver Version: 525.105.17 CUDA Version: 12.1

Reproduction Steps:
step 1 : git clone git@github.com:Eddie-Wang1120/Eddie-Wang-Hackathon2023.git
step 2 : 进入examples/whisper目录
step 3 : 将build.py 304行开始的以下代码注释,将275行开始的相同代码Uncomment。
image
step 4 : bash run_int8_kv_cache.sh

Expected Behavior:
应该出现的正确结果如下:
image

Actual Behavior:
实际出现的结果如下:
image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant