You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @spoonbobo , not yet. We're currently working on a general backend for structures like encoder-decoder and multimodal models. Encoder-decoder work is in progress and multimodal follows it. The progress is tracked in NVIDIA/TensorRT-LLM#800.
Meanwhie, if you're referring to a Triton Python backend, do you think it's ok for users to implement a multimodal workflow based on the gpt example
hi @symphonylyh. Appreciated the efforts you've put on providing general encoder-decoder support. Haven't tried implement a workflow based on this example, I think definitely worth a try.
Does tensorrtllm_backend supports multimodal LLM like LLaVA like those listed in https://github.com/NVIDIA/TensorRT-LLM/tree/main/examples/multimodal?
The text was updated successfully, but these errors were encountered: