diff --git a/CMakeLists.txt b/CMakeLists.txt index 3a1fc3d22..6bb9260ae 100644 --- a/CMakeLists.txt +++ b/CMakeLists.txt @@ -164,7 +164,7 @@ if(NOT TRITON_CORE_HEADERS_ONLY) option(TRITON_ENABLE_NVTX "Include NVTX support in server" OFF) option(TRITON_ENABLE_GPU "Enable GPU support in server" ON) option(TRITON_ENABLE_MALI_GPU "Enable Arm Mali GPU support in server" OFF) - set(TRITON_MIN_COMPUTE_CAPABILITY "6.0" CACHE STRING + set(TRITON_MIN_COMPUTE_CAPABILITY "7.5" CACHE STRING "The minimum CUDA compute capability supported by Triton" ) set(TRITON_EXTRA_LIB_PATHS "" CACHE PATH "Extra library paths for Triton Server build") diff --git a/src/ensemble_scheduler/ensemble_scheduler.cc b/src/ensemble_scheduler/ensemble_scheduler.cc index b16567dd7..609c1ff28 100644 --- a/src/ensemble_scheduler/ensemble_scheduler.cc +++ b/src/ensemble_scheduler/ensemble_scheduler.cc @@ -1200,11 +1200,6 @@ EnsembleContext::CheckAndSetEnsembleOutput( ready = false; break; } - // check if the output is provided - else if (tensor[iteration_count].data_ == nullptr) { - ready = false; - break; - } } } if (!ready) { @@ -1224,6 +1219,12 @@ EnsembleContext::CheckAndSetEnsembleOutput( auto& tensor_data = tensor_data_[output_pair.first]; auto& tensor = tensor_data.tensor_[iteration_count]; + if (tensor.data_ == nullptr) { + LOG_VERBOSE(1) << "Composing models did not output tensor " + << output_pair.first; + continue; + } + auto shape = ReshapeTensorDims( output_pair.second, (lrequest->BatchSize() != 0), tensor_data.batch_size_, tensor.data_->OriginalShape());