Skip to content

Commit

Permalink
Update after review
Browse files Browse the repository at this point in the history
Review comments:

- foundation-model-stack#52 (review)

Signed-off-by: Martin Hickey <[email protected]>
  • Loading branch information
hickeyma committed Feb 26, 2024
1 parent e898b1d commit 647da3e
Show file tree
Hide file tree
Showing 3 changed files with 5 additions and 2 deletions.
2 changes: 2 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,8 @@ pip install -U datasets
pip install -e .
```

> Note: If you wish to use [FlashAttention](https://github.com/Dao-AILab/flash-attention), then you need to install these requirements: `pip install -r flashattn_requirements`. [FlashAttention](https://github.com/Dao-AILab/flash-attention) requires the [CUDA Toolit](https://developer.nvidia.com/cuda-toolkit) to be pre-installed.
## Data format
The data format expectation is a single column text. The trainer is configured to expect a response template as a string. For example, if one wants to prepare the `alpaca` format data to feed into this trainer, it is quite easy and can be done with the following code.

Expand Down
2 changes: 0 additions & 2 deletions torch_requirements.txt → flashattn_requirements.txt
Original file line number Diff line number Diff line change
@@ -1,4 +1,2 @@
packaging
torch
wheel
flash-attn
3 changes: 3 additions & 0 deletions requirements.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,8 @@
numpy
accelerate>=0.20.3
packaging
transformers>=4.34.1
torch
aim==3.17.5
sentencepiece
tokenizers>=0.13.3
Expand All @@ -10,3 +12,4 @@ ninja
peft>=0.8.0
datasets>=2.15.0
fire

0 comments on commit 647da3e

Please sign in to comment.