Skip to content

Commit

Permalink
Add all sizes of LLaMA on Together (#1740)
Browse files Browse the repository at this point in the history
  • Loading branch information
yifanmai authored Jul 25, 2023
1 parent 87ee885 commit 33bda52
Show file tree
Hide file tree
Showing 5 changed files with 52 additions and 7 deletions.
23 changes: 22 additions & 1 deletion src/helm/benchmark/static/schema.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -407,13 +407,34 @@ models:
num_parameters: 30000000000
release_date: 2022-11-15
todo: true
- name: huggingface/llama-7b
- name: meta/llama-7b
display_name: LLaMA (7B)
description: LLaMA is a collection of foundation language models ranging from 7B to 65B parameters.
creator_organization: Meta
access: open
num_parameters: 7000000000
release_date: 2023-02-24
- name: meta/llama-13b
display_name: LLaMA (13B)
description: LLaMA is a collection of foundation language models ranging from 7B to 65B parameters.
creator_organization: Meta
access: open
num_parameters: 13000000000
release_date: 2023-02-24
- name: meta/llama-30b
display_name: LLaMA (30B)
description: LLaMA is a collection of foundation language models ranging from 7B to 65B parameters.
creator_organization: Meta
access: open
num_parameters: 30000000000
release_date: 2023-02-24
- name: meta/llama-65b
display_name: LLaMA (65B)
description: LLaMA is a collection of foundation language models ranging from 7B to 65B parameters.
creator_organization: Meta
access: open
num_parameters: 65000000000
release_date: 2023-02-24
- name: meta/llama-2-7b
display_name: LLaMA-2 (7B)
description: Llama 2 pretrained models are trained on 2 trillion tokens, and have double the context length than Llama 1.
Expand Down
5 changes: 4 additions & 1 deletion src/helm/benchmark/window_services/window_service_factory.py
Original file line number Diff line number Diff line change
Expand Up @@ -172,7 +172,10 @@ def get_window_service(model_name: str, service: TokenizerService) -> WindowServ
elif model_name == "nvidia/megatron-gpt2":
window_service = MegatronWindowService(service)
elif model_name in [
"together/llama-7b",
"meta/llama-7b",
"meta/llama-13b",
"meta/llama-30b",
"meta/llama-65b",
"together/alpaca-7b",
"together/vicuna-13b",
]:
Expand Down
4 changes: 2 additions & 2 deletions src/helm/proxy/clients/test_together_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,7 @@ def teardown_method(self, method):
),
(
Request(
model="together/llama-7b",
model="meta/llama-7b",
prompt="I am a computer scientist.",
temperature=0,
num_completions=4,
Expand All @@ -57,7 +57,7 @@ def teardown_method(self, method):
"echo": True,
"logprobs": 3,
"max_tokens": 24,
"model": "llama-7b-full-precision",
"model": "huggyllama/llama-7b",
"n": 4,
"prompt": "I am a computer scientist.",
"request_type": "language-model-inference",
Expand Down
10 changes: 8 additions & 2 deletions src/helm/proxy/clients/together_client.py
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@
_ASYNC_MODELS: Set[str] = {
# Legacy models
"alpaca-7b",
"llama-7b",
"pythia-7b",
"vicuna-13b",
# Production models
Expand All @@ -26,6 +25,10 @@
"dolly-v2-3b",
"dolly-v2-7b",
"dolly-v2-12b",
"llama-7b",
"llama-13b",
"llama-30b",
"llama-65b",
"llama-2-7b",
"llama-2-13b",
"llama-2-70b",
Expand All @@ -52,7 +55,6 @@
# alpaca-7b is half-precision
# alpaca-7b-full-precision is full-precision
"alpaca-7b": "alpaca-7b-full-precision",
"llama-7b": "llama-7b-full-precision",
"pythia-7b": "pythia-7b-full-precision",
"vicuna-13b": "vicuna-13b-full-precision",
# Production models
Expand All @@ -61,6 +63,10 @@
"dolly-v2-3b": "databricks/dolly-v2-3b",
"dolly-v2-7b": "databricks/dolly-v2-7b",
"dolly-v2-12b": "databricks/dolly-v2-12b",
"llama-7b": "huggyllama/llama-7b",
"llama-13b": "huggyllama/llama-13b",
"llama-30b": "huggyllama/llama-30b",
"llama-65b": "huggyllama/llama-65b",
"llama-2-7b": "togethercomputer/llama-2-7b",
"llama-2-13b": "togethercomputer/llama-2-13b",
"llama-2-70b": "togethercomputer/llama-2-70b",
Expand Down
17 changes: 16 additions & 1 deletion src/helm/proxy/models.py
Original file line number Diff line number Diff line change
Expand Up @@ -301,7 +301,22 @@ def engine(self) -> str:
# Meta
Model(
group="together",
name="together/llama-7b",
name="meta/llama-7b",
tags=[TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG],
),
Model(
group="together",
name="meta/llama-13b",
tags=[TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG],
),
Model(
group="together",
name="meta/llama-30b",
tags=[TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG],
),
Model(
group="together",
name="meta/llama-65b",
tags=[TEXT_MODEL_TAG, FULL_FUNCTIONALITY_TEXT_MODEL_TAG],
),
Model(
Expand Down

0 comments on commit 33bda52

Please sign in to comment.