Skip to content

Issues: NVIDIA/TensorRT-LLM

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Assignee
Filter by who’s assigned
Sort

Issues list

KV cache within DeepSeekV3
#2738 opened Feb 3, 2025 by usajid14
Lora error while building tensorrt llm engine for mllama bug Something isn't working
#2733 opened Feb 1, 2025 by nbowon
2 of 4 tasks
Unable to Install tensorrt_llm 17 Due to Flashinfer Git Clone Failure bug Something isn't working
#2729 opened Jan 31, 2025 by ValeGian
2 of 4 tasks
"Trying to remove block n by 0 that is not in hash map" spam in release 0.17 bug Something isn't working
#2727 opened Jan 31, 2025 by aikitoria
2 of 4 tasks
Whisper example not returning transcription in orig language bug Something isn't working
#2721 opened Jan 27, 2025 by haiderasad
2 of 4 tasks
When to expect new development versions triaged Issue has been triaged by maintainers
#2720 opened Jan 24, 2025 by ttim
Memory Leak: Tensorrt-llm process
#2716 opened Jan 24, 2025 by payingguest
Are multimodal models supported by trtllm-serve? OpenAI API triaged Issue has been triaged by maintainers
#2714 opened Jan 23, 2025 by xiaoyuzju
how to compile deepseekv3 ? Installation triaged Issue has been triaged by maintainers
#2711 opened Jan 22, 2025 by zmtttt
Support for Blackwell and Thor triaged Issue has been triaged by maintainers
#2710 opened Jan 21, 2025 by phantaurus
NVILA support (Qwen2) bug Something isn't working
#2707 opened Jan 21, 2025 by danigarciaoca
convert NVILA with 0.16.0 bug Something isn't working Investigating LLM API/Workflow triaged Issue has been triaged by maintainers
#2706 opened Jan 20, 2025 by dzy130120
2 of 4 tasks
Support for int2/int3 quantization Investigating Low Precision Issue about lower bit quantization, including int8, int4, fp8 triaged Issue has been triaged by maintainers
#2704 opened Jan 20, 2025 by ZHITENGLI
quantized model using AWQ and lora weights Investigating Low Precision Issue about lower bit quantization, including int8, int4, fp8 triaged Issue has been triaged by maintainers
#2703 opened Jan 17, 2025 by shuyuan-wang
Wrong outputs with FP8 kv_cache reuse bug Something isn't working Investigating KV-Cache Management triaged Issue has been triaged by maintainers
#2699 opened Jan 16, 2025 by lishicheng1996
2 of 4 tasks
ProTip! What’s not been updated in a month: updated:<2025-01-03.