Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix inference container #135

Merged
merged 25 commits into from
Mar 17, 2022
Merged

Fix inference container #135

merged 25 commits into from
Mar 17, 2022

Conversation

albert17
Copy link
Contributor

@albert17 albert17 commented Mar 9, 2022

No description provided.

@albert17
Copy link
Contributor Author

albert17 commented Mar 9, 2022

Comment on lines 57 to 67
# Triton Server
FROM ${FULL_IMAGE} as full
WORKDIR /opt/tritonserver
COPY --chown=1000:1000 --from=full /opt/tritonserver/LICENSE .
COPY --chown=1000:1000 --from=full /opt/tritonserver/TRITON_VERSION .
COPY --chown=1000:1000 --from=full /opt/tritonserver/NVIDIA_Deep_Learning_Container_License.pdf .
COPY --chown=1000:1000 --from=full /opt/tritonserver/bin bin/
COPY --chown=1000:1000 --from=full /opt/tritonserver/lib lib/
COPY --chown=1000:1000 --from=full /opt/tritonserver/include include/
COPY --chown=1000:1000 --from=full /opt/tritonserver/repoagents/ repoagents/
COPY --chown=1000:1000 --from=full /usr/bin/serve /usr/bin/.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we will need to add these lines also in dockerfile.torch.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@IamGianluca Triton server is already installed in pytorch and tensorflow inference containers

albertoa@pursuit-dgxstation:~/Projects/Merlin/docker/inference$ docker run --pull always --gpus=all -it --ipc=host --cap-add SYS_NICE nvcr.io/nvstaging/merlin/merlin-pytorch-inference:22.03 /bin/bash
22.03: Pulling from nvstaging/merlin/merlin-pytorch-inference
Digest: sha256:8be045dfbb42ea128aca833b78f4847bcc69557ba97f5e65a3703f50606fc646
Status: Image is up to date for nvcr.io/nvstaging/merlin/merlin-pytorch-inference:22.03

=============================
== Triton Inference Server ==
=============================

NVIDIA Release 22.02 (build 32400308)

Copyright (c) 2018-2021, NVIDIA CORPORATION & AFFILIATES.  All rights reserved.

Various files include modifications (c) NVIDIA CORPORATION.  All rights reserved.

This container image and its contents are governed by the NVIDIA Deep Learning Container License.
By pulling and using the container, you accept the terms and conditions of this license:
https://developer.nvidia.com/ngc/nvidia-deep-learning-container-license

NOTE: Legacy NVIDIA Driver detected.  Compatibility mode ENABLED.

root@3f045a14f399:/opt/tritonserver# ls -la
total 12396
drwxr-xr-x  1 root          root             4096 Mar  4 10:17 .
drwxr-xr-x  1 root          root             4096 Feb 18 05:47 ..
-rw-rw-r--  1 triton-server triton-server    1485 Feb 18 01:23 LICENSE
-rw-rw-r--  1 triton-server triton-server 3012640 Feb 18 01:23 NVIDIA_Deep_Learning_Container_License.pdf
-rw-rw-r--  1 triton-server triton-server       7 Feb 18 01:23 TRITON_VERSION
drwxr-xr-x  1 triton-server triton-server    4096 Mar  4 10:17 backends
drwxr-xr-x  2 triton-server triton-server    4096 Feb 18 05:47 bin
drwxrwxr-x 15 root          root             4096 Mar  4 08:20 cmake-3.21.1
-rw-r--r--  1 root          root          9629567 Jul 27  2021 cmake-3.21.1.tar.gz
drwxr-xr-x  3 triton-server triton-server    4096 Feb 18 05:47 include
drwxr-xr-x  2 triton-server triton-server    4096 Feb 18 05:47 lib
-rwxrwxr-x  1 triton-server triton-server    4266 Feb 18 05:41 nvidia_entrypoint.sh
drwxr-xr-x  1 triton-server triton-server    4096 Feb 18 05:48 repoagents
root@3f045a14f399:/opt/tritonserver# ls -la bin/
total 10848
drwxr-xr-x 2 triton-server triton-server     4096 Feb 18 05:47 .
drwxr-xr-x 1 root          root              4096 Mar  4 10:17 ..
-rwxr-xr-x 1 triton-server triton-server 11092616 Feb 18 01:38 tritonserver
root@3f045a14f399:/opt/tritonserver#

This problems only happens in ctr (hugectr) since the base image is very minimal. I was working in getting smaller inference container since size was a problem for cloud providers, and I cut too much.

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Got it! Thank you for the explanation @albert17 👍

@yingcanw
Copy link
Contributor

@albert17 ,when this PR can be merged, because inference-related CI of HugeCTR backend has failed to pass using old container

@albert17 albert17 linked an issue Mar 15, 2022 that may be closed by this pull request
@albert17
Copy link
Contributor Author

merling-pytorch-inference and merlin-tensorflow-inference nightly are pushed

@rnyak
Copy link
Contributor

rnyak commented Mar 16, 2022

@albert17 did you add FIL to both tensorflow-inference and pytorch-inference nightly containers?

@albert17
Copy link
Contributor Author

@IamGianluca Please try

docker pull nvcr.io/nvidia/merlin/merlin-inference:nightly

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
5 participants