-
-
Notifications
You must be signed in to change notification settings - Fork 57
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Switch from Tensorflow Serving to Triton #622
Labels
Comments
I am interested to work on this issue , if no one is currently working on it |
teolemon
moved this to To discuss and validate
in 🤖 Artificial Intelligence @ Open Food Facts
Sep 20, 2022
raphael0202
added a commit
that referenced
this issue
Oct 28, 2022
- switch to Triton, use gRPC protocol for faster transfer - only request needed output fields (resolves #942) - get rid of CategoryIndex and protobuf-generated files
raphael0202
added a commit
that referenced
this issue
Oct 31, 2022
- switch to Triton, use gRPC protocol for faster transfer - only request needed output fields (resolves #942) - get rid of CategoryIndex and protobuf-generated files
raphael0202
added a commit
that referenced
this issue
Nov 8, 2022
- switch to Triton, use gRPC protocol for faster transfer - only request needed output fields (resolves #942) - get rid of CategoryIndex and protobuf-generated files
raphael0202
added a commit
that referenced
this issue
Nov 8, 2022
- switch to Triton, use gRPC protocol for faster transfer - only request needed output fields (resolves #942) - get rid of CategoryIndex and protobuf-generated files
raphael0202
added a commit
that referenced
this issue
Nov 8, 2022
- switch to Triton, use gRPC protocol for faster transfer - only request needed output fields (resolves #942) - get rid of CategoryIndex and protobuf-generated files
raphael0202
added a commit
that referenced
this issue
Nov 8, 2022
- switch to Triton, use gRPC protocol for faster transfer - only request needed output fields (resolves #942) - get rid of CategoryIndex and protobuf-generated files
The switch to Triton is done in preprod, closing this issue. |
Repository owner
moved this from To discuss and validate
to Done
in 🤖 Artificial Intelligence @ Open Food Facts
Nov 8, 2022
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
We currently use Tensorflow Serving to serve all models:
However, Tensorflow Serving only supports the SavedModel format, which is specific to Tensorflow. It means we cannot easily integrate models trained with other frameworks (pytorch, scikit-learn,...).
ONNX is a standard format to save model architecture and weights supported by all major frameworks, including Tensorflow, Pytorch and scikit-learn.
Proposal
Test Triton, an open-source inference server developed by Nvidia. Supported formats include TF SavedModel and ONNX. Runtime on both GPU and CPU.
If tests are successful, switch from Tensorflow Serving to Triton server.
The text was updated successfully, but these errors were encountered: