Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Do not merge!] Build: Remove TRT model generation for V100 #7712

Draft
wants to merge 3,421 commits into
base: r24.10
Choose a base branch
from

Conversation

pvijayakrish
Copy link
Contributor

What does the PR do?

Remove TRT model generation on V100.

Checklist

  • PR title reflects the change and is of format <commit_type>: <Title>
  • Changes are described in the pull request.
  • Related issues are referenced.
  • Populated github labels field
  • Added test plan and verified test passes.
  • Verified that the PR passes existing CI.
  • Verified copyright is correct on all changed files.
  • Added succinct git squash message before merging ref.
  • All template sections are filled out.
  • Optional: Additional screenshots for behavior/output changes with before/after.

Commit Type:

Check the conventional commit type
box here and add the label to the github PR.

  • build
  • ci
  • docs
  • feat
  • fix
  • perf
  • refactor
  • revert
  • style
  • test

rmccorm4 and others added 30 commits November 15, 2023 19:57
* name changes

* updated names
* Add documentation on generative sequence

* Address comment

* Reflect the "iterative" change
…#6620)

* Extend request objects lifetime

* Remove explicit TRITONSERVER_InferenceRequestDelete

* Format fix

* Include the inference_request_ initialization to cover RequestNew

---------

Co-authored-by: Neelay Shah <[email protected]>
This fixes the issue where python client has
`AttributeError: 'NoneType' object has no attribute 'enum_types_by_name'
errors after python version is updated.
* Update README and versions for 2.40.0 / 23.11 (#6544)

* Removing path construction to use SymLink alternatives

* Update version for PyTorch

* Update windows Dockerfile configuration

* Update triton version to 23.11

* Update README and versions for 2.40.0 / 23.11

* Fix typo

* Ading 'ldconfig' to configure dynamic linking in container (#6602)

* Point to tekit_backend (#6616)

* Point to tekit_backend

* Update version

* Revert tekit changes (#6640)

---------

Co-authored-by: Kris Hung <[email protected]>
* New testing to confirm large request timeout values can be passed and retrieved within Python BLS models.
* Add test for optional internal tensor within an ensemble

* Fix up
* Set CMake version to 3.27.7

* Set CMake version to 3.27.7

* Fix double slash typo
* Mlflow plugin fix
* Unify iGPU test build with x86 ARM

* adding TRITON_IGPU_BUILD to core build definition; adding logic to skip caffe2plan test if TRITON_IGPU_BUILD=1

* re-organizing some copies in Dockerfile.QA to fix igpu devel build

* Pre-commit fix

---------

Co-authored-by: kyle <[email protected]>
* adding default value for TRITON_IGPU_BUILD=OFF

* fix newline

---------

Co-authored-by: kyle <[email protected]>
* Add test case for decoupled model raising exception

* Remove unused import

* Address comment
@pvijayakrish pvijayakrish requested a review from mc-nv October 16, 2024 22:04
@pvijayakrish pvijayakrish changed the title Build: Remove TRT model generation for V100 [Do not merge!] Build: Remove TRT model generation for V100 Oct 16, 2024
@@ -512,7 +512,7 @@ chmod -R 777 $FORMATDESTDIR
python3 $SRCDIR/gen_qa_trt_data_dependent_shape.py --models_dir=$DATADEPENDENTDIR
chmod -R 777 $DATADEPENDENTDIR
# Make shared library for custom Hardmax plugin.
(git clone -b release/10.0 https://github.com/NVIDIA/TensorRT.git && \
(git clone -b release/10.5 https://github.com/NVIDIA/TensorRT.git && \
Copy link
Contributor

@mc-nv mc-nv Oct 16, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Just to avoid us from managing the version

Suggested change
(git clone -b release/10.5 https://github.com/NVIDIA/TensorRT.git && \
docker pull ${TENSORRT_IMAGE}
TENSORRT_VERSION="$(docker inspect ${TENSORRT_IMAGE} --format '{{index .Config.Labels "com.nvidia.tensorrt.version"}}' | cut -d . -f -2)"
(git clone -b release/${TENSORRT_VERSION} https://github.com/NVIDIA/TensorRT.git && \

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @mc-nv! Ack the suggestions. This is not to be merged and is still under development and is only a work around for generating models on V100 to unblock testing.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Development

Successfully merging this pull request may close these issues.