fix: Fix the server runtime errors on cpu only platform and with pytorch backend #8272
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What does the PR do?
The server docker image built for cpu only platform and with pytorch backend was failing to run due to the missing cuda library stubs. This commit is adding those libraries to the build script.
Checklist
Agreement
<commit_type>: <Title>
pre-commit install, pre-commit run --all
)Commit Type:
Check the conventional commit type
box here and add the label to the github PR.
Related PRs:
None
Where should the reviewer start?
build.py
Test plan:
build cpu only container with pytorch backend using the following command
python3 build.py --enable-logging --endpoint http --backend pytorch --image=gpu-base,nvcr.io/nvidia/tritonserver:25.06-py3-min
run the container to serve any pytorch model
docker run --rm -p8000:8000 -p8001:8001 -p8002:8002 -v/home/ubuntu/server/pt_model:/models tritonserver:latest tritonserver --model-repository=/models
With this fix, the container starts successfully and serves the model
Caveats:
None
Background
Related Issues: (use one of the action keywords Closes / Fixes / Resolves / Relates to)