You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi, I've been trying to deploy an mtf model to the NVIDIA Triton Inference Server by converting the SavedModel (output of model.export()) to an onnx file with no luck. I've been receiving several errors, but the main recurring one seems to be regarding the lack of a registered SentencePieceOps. Is there a tutorial for an mtf model deployment available anywhere?
The text was updated successfully, but these errors were encountered:
Hi! Thank you for replying. I am using a language model - T5 to be specific. I've been using the mtf model implementation from Google-research's GitHub. I managed to convert it to a Pytorch model, and I've been successful at converting to onnx.
Hi, I've been trying to deploy an mtf model to the NVIDIA Triton Inference Server by converting the SavedModel (output of model.export()) to an onnx file with no luck. I've been receiving several errors, but the main recurring one seems to be regarding the lack of a registered SentencePieceOps. Is there a tutorial for an mtf model deployment available anywhere?
The text was updated successfully, but these errors were encountered: