You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jan 21, 2025. It is now read-only.
Hi, I've been trying to deploy an mtf model to the NVIDIA Triton Inference Server by converting the SavedModel (output of model.export()) to an onnx file with no luck. I've been receiving several errors, but the main recurring one seems to be regarding the lack of a registered SentencePieceOps. Is there a tutorial for an mtf model deployment available anywhere?
The text was updated successfully, but these errors were encountered:
Hi! Thank you for replying. I am using a language model - T5 to be specific. I've been using the mtf model implementation from Google-research's GitHub. I managed to convert it to a Pytorch model, and I've been successful at converting to onnx.
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hi, I've been trying to deploy an mtf model to the NVIDIA Triton Inference Server by converting the SavedModel (output of model.export()) to an onnx file with no luck. I've been receiving several errors, but the main recurring one seems to be regarding the lack of a registered SentencePieceOps. Is there a tutorial for an mtf model deployment available anywhere?
The text was updated successfully, but these errors were encountered: