Skip to content

Commit

Permalink
Update installation instructions to prevent/resolve nccl error, updat…
Browse files Browse the repository at this point in the history
…e poetry lock to vllm 0.4.2, update flash attention instructions for vllm 0.4.2
  • Loading branch information
XkunW committed May 23, 2024
1 parent 3ec9069 commit 9e532c4
Show file tree
Hide file tree
Showing 3 changed files with 198 additions and 57 deletions.
1 change: 1 addition & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -23,6 +23,7 @@ pip install wheel
export CUDA_HOME=/pkgs/cuda-12.3

pip install flash-attn --no-build-isolation
pip install vllm-flash-attn
```

## Launch an inference server
Expand Down
Loading

0 comments on commit 9e532c4

Please sign in to comment.