Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BUG] tritonserver not found in Merlin-inference 22.03 image #373

Closed
IamGianluca opened this issue Mar 5, 2022 · 7 comments
Closed

[BUG] tritonserver not found in Merlin-inference 22.03 image #373

IamGianluca opened this issue Mar 5, 2022 · 7 comments
Labels
bug Something isn't working status/needs-triage

Comments

@IamGianluca
Copy link

Hi,

It seems that tritonserver is not installed in the latest stable merlin-inference image (22.03).

==========
== CUDA ==
==========

NVIDIA Release  (build )
CUDA Version 11.6.0.021

Container image Copyright (c) 2022, NVIDIA CORPORATION & AFFILIATES. All rights reserved.

Various files include modifications (c) NVIDIA CORPORATION & AFFILIATES.  All rights reserved.

This container image and its contents are governed by the NVIDIA Deep Learning Container License.
By pulling and using the container, you accept the terms and conditions of this license:
https://developer.nvidia.com/ngc/nvidia-deep-learning-container-license

NOTE: MOFED driver for multi-node communication was not detected.
      Multi-node communication performance may be reduced.

root@746a74b9ea7b:/# tritonserver --model-repository=/root/models/ --model-control-mode=explicit
bash: tritonserver: command not found
@IamGianluca IamGianluca added bug Something isn't working status/needs-triage labels Mar 5, 2022
@albert17
Copy link
Contributor

albert17 commented Mar 9, 2022

HI @IamGianluca looking into this right now

@albert17
Copy link
Contributor

albert17 commented Mar 9, 2022

Found the issue. Working on fixing this issue

@IamGianluca
Copy link
Author

Thank you for looking into that @albert17 🙏

@albert17
Copy link
Contributor

albert17 commented Mar 9, 2022

Created a PR: NVIDIA-Merlin/Merlin#135

@rnyak
Copy link
Contributor

rnyak commented Mar 10, 2022

@IamGianluca you can pull nvcr.io/nvidia/merlin/merlin-pytorch-inference:22.03 image as well. You might need to install torch and torchmetrics in the container.

@IamGianluca
Copy link
Author

IamGianluca commented Mar 10, 2022

Good point @rnyak. I should probably use nvcr.io/nvidia/merlin/merlin-pytorch-inference:22.03 since I'm developing in pytorch. I'll try that. What is the reasoning behind not installing pytorch and torchmetrics in that image?

@EvenOldridge
Copy link
Member

Hi @IamGianluc. Our inference container uses Triton inference server which can serve a pytorch model without requiring the library. We're going to release unified training and inference containers in the next release which should make this unnecessary.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working status/needs-triage
Projects
None yet
Development

No branches or pull requests

4 participants