Skip to content

Commit

Permalink
delete cache
Browse files Browse the repository at this point in the history
  • Loading branch information
IlyasMoutawwakil committed Dec 10, 2024
1 parent 07218dc commit 91e95cd
Show file tree
Hide file tree
Showing 4 changed files with 6 additions and 7 deletions.
4 changes: 3 additions & 1 deletion .github/workflows/test_cli_cuda_tensorrt_llm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,9 @@ jobs:
contains( github.event.pull_request.labels.*.name, 'examples')
}}
name: Run examples
run: pytest tests/test_examples.py -x -s -k "cli and cuda and trt"
run: |
huggingface-cli delete-cache
pytest tests/test_examples.py -x -s -k "cli and cuda and trt"
cli_cuda_tensorrt_llm_multi_gpu_tests:
if: ${{
Expand Down
7 changes: 3 additions & 4 deletions examples/cuda_trt_llama.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -15,11 +15,10 @@ launcher:
backend:
device: cuda
device_ids: 0
force_export: true
model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
max_prompt_length: 64
max_new_tokens: 32
max_batch_size: 4
max_new_tokens: 32
max_prompt_length: 64
model: TinyLlama/TinyLlama-1.1B-Chat-v1.0

scenario:
input_shapes:
Expand Down
1 change: 0 additions & 1 deletion optimum_benchmark/backends/tensorrt_llm/backend.py
Original file line number Diff line number Diff line change
Expand Up @@ -46,7 +46,6 @@ def load_trtmodel_from_pretrained(self) -> None:
max_batch_size=self.config.max_batch_size,
max_new_tokens=self.config.max_new_tokens,
max_beam_width=self.config.max_beam_width,
force_export=self.config.force_export,
**self.config.model_kwargs,
)

Expand Down
1 change: 0 additions & 1 deletion optimum_benchmark/backends/tensorrt_llm/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,6 @@ class TRTLLMConfig(BackendConfig):
pp: int = 1
use_fp8: bool = False
dtype: str = "float16"
force_export: bool = False
optimization_level: int = 2
use_cuda_graph: bool = False

Expand Down

0 comments on commit 91e95cd

Please sign in to comment.