Skip to content

Commit

Permalink
Apply suggestions from code review
Browse files Browse the repository at this point in the history
Co-authored-by: dyastremsky <[email protected]>
  • Loading branch information
oandreeva-nv and dyastremsky authored Oct 31, 2023
1 parent 04dca66 commit e3295c9
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion Quick_Deploy/vLLM/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -104,7 +104,7 @@ To use this container to launch Triton, you can use the docker command below.
```
docker run --gpus all -it --net=host --rm -p 8001:8001 --shm-size=1G --ulimit memlock=-1 --ulimit stack=67108864 -v ${PWD}:/work -w /work nvcr.io/nvidia/tritonserver:<xx.yy>-vllm-python-py3 tritonserver --model-store ./model_repository
```
Here and later throughout the tutorial \<xx.yy\> is the version of Triton
Throughout the tutorial, \<xx.yy\> is the version of Triton
that you want to use (and pulled above). Please note, that Triton's vLLM
container was first published in 23.10 release, so any prior version
will not work.
Expand Down

0 comments on commit e3295c9

Please sign in to comment.