Skip to content

Commit

Permalink
update distributed configs
Browse files Browse the repository at this point in the history
  • Loading branch information
IlyasMoutawwakil committed Nov 28, 2023
1 parent c27c708 commit f00d042
Show file tree
Hide file tree
Showing 4 changed files with 22 additions and 12 deletions.
5 changes: 1 addition & 4 deletions examples/running-llamas/configs/fp16+dp=2.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -7,10 +7,7 @@ experiment_name: fp16+dp=2

launcher:
nproc_per_node: 2

backend:
# only serves for init torch.distributed
deepspeed_inference: true
rdzv_endpoint: localhost:29511

hydra:
job:
Expand Down
13 changes: 10 additions & 3 deletions examples/running-llamas/configs/fp16+gptq+exllamav1+dp=2.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,13 @@
defaults:
- _base_
- _self_
- override launcher: process
- override launcher: torchrun

experiment_name: fp16+gptq+exllamav1
experiment_name: fp16+gptq+exllamav1+dp=2

launcher:
nproc_per_node: 2
rdzv_endpoint: localhost:29522

backend:
quantization_scheme: gptq
Expand All @@ -12,6 +16,9 @@ backend:
version: 1

hydra:
job:
env_set:
CUDA_VISIBLE_DEVICES: 0,1
sweeper:
params:
model: TheBloke/LLaMa-7B-GPTQ,TheBloke/LLaMa-13B-GPTQ
model: TheBloke/LLaMa-7B-GPTQ,TheBloke/LLaMa-13B-GPTQ,TheBloke/LLaMa-70B-GPTQ
13 changes: 10 additions & 3 deletions examples/running-llamas/configs/fp16+gptq+exllamav2+dp=2.yaml
Original file line number Diff line number Diff line change
@@ -1,9 +1,13 @@
defaults:
- _base_
- _self_
- override launcher: process
- override launcher: torchrun

experiment_name: fp16+gptq+exllamav2
experiment_name: fp16+gptq+exllamav2+dp=2

launcher:
nproc_per_node: 2
rdzv_endpoint: localhost:29533

backend:
# for some reason core gets dumped
Expand All @@ -15,6 +19,9 @@ backend:
version: 2

hydra:
job:
env_set:
CUDA_VISIBLE_DEVICES: 0,1
sweeper:
params:
model: TheBloke/LLaMa-7B-GPTQ,TheBloke/LLaMa-13B-GPTQ
model: TheBloke/LLaMa-7B-GPTQ,TheBloke/LLaMa-13B-GPTQ,TheBloke/LLaMa-70B-GPTQ
3 changes: 1 addition & 2 deletions examples/running-llamas/configs/fp16+tp=2.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,9 +6,8 @@ defaults:
experiment_name: fp16+tp=2

launcher:
# this is not needed since we have a resolver
# that takes len(cuda_devices) but just to be verbose
nproc_per_node: 2
rdzv_endpoint: localhost:29544

backend:
deepspeed_inference: true
Expand Down

0 comments on commit f00d042

Please sign in to comment.