Skip to content

Commit

Permalink
Merge branch 'main' into davide/fix-docs-minio
Browse files Browse the repository at this point in the history
  • Loading branch information
aittalam authored Jan 17, 2025
2 parents 4854e58 + 5b4275b commit 74497f6
Show file tree
Hide file tree
Showing 7 changed files with 41 additions and 190 deletions.
37 changes: 1 addition & 36 deletions docs/source/get-started/suggested-models.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ running instance. To get a list of suggested models, you can use the following c
```console
user@host:~/lumigator$ curl -s http://localhost:8000/api/v1/models/summarization | jq
{
"total": 9,
"total": 7,
"items": [
{
"name": "facebook/bart-large-cnn",
Expand Down Expand Up @@ -83,8 +83,6 @@ launched it.
| Model Type | Model | HuggingFace | API | llamafile |
|------------|------------------------------------------|-------------|-----|-----------|
| seq2seq | facebook/bart-large-cnn | X | | |
| seq2seq | longformer-qmsum-meeting-summarization | X | | |
| seq2seq | mrm8488/t5-base-finetuned-summarize-news | X | | |
| seq2seq | Falconsai/text_summarization | X | | |
| causal | gpt-4o-mini, gpt-4o | | X | |
| causal | open-mistral-7b | | X | |
Expand All @@ -110,39 +108,6 @@ evaluation are:
| `no_repeat_ngram_size` | All n-grams of that size can only occur once | 3 |
| `num_beams` | Number of beams for beam search | 4 |

## Longformer QMSum Meeting Summarization

The [`longformer-qmsum-meeting-summarization`](https://huggingface.co/mikeadimech/longformer-qmsum-meeting-summarization)
model is a fine-tuned version of [alenai/led-base-16384](https://huggingface.co/allenai/led-base-16384)
for summarization.

As described in [Longformer: The Long-Document Transformer](https://arxiv.org/pdf/2004.05150.pdf) by
Iz Beltagy, Matthew E. Peters, Arman Cohan, `led-base-16384` was initialized from `bart-base` since
both models share the exact same architecture, but modified for long-range summarization and
question answering.

The model has 162M parameters (FP32), and the model size is 648MB. There are no
summarization-specific parameters for this model.

## T5 Base Finetuned Summarize News

The [`mrm8488/t5-base-finetuned-summarize-news`](https://huggingface.co/mrm8488/t5-base-finetuned-summarize-news)
model is a [Google's T5](https://ai.googleblog.com/2020/02/exploring-transfer-learning-with-t5.html)
base fine-tuned on [News Summary](https://www.kaggle.com/sunnysai12345/news-summary) dataset for
summarization downstream task.

The model has 223M parameters (FP32), and the model size is 892MB. The default parameters used for
evaluation are:

| Parameter Name | Description | Value |
|------------------------|--------------------------------------------------------|-------|
| `max_length` | Maximum length of the summary | 200 |
| `min_length` | Minimum length of the summary | 30 |
| `length_penalty` | Length penalty to apply during beam search | 2.0 |
| `early_stopping` | Controls the stopping condition for beam-based methods | true |
| `no_repeat_ngram_size` | All n-grams of that size can only occur once | 3 |
| `num_beams` | Number of beams for beam search | 4 |

## Falconsai Text Summarization

The [`Falconsai/text_summarization`](https://huggingface.co/Falconsai/text_summarization) model is
Expand Down
2 changes: 0 additions & 2 deletions lumigator/python/mzai/backend/backend/config_templates.py
Original file line number Diff line number Diff line change
Expand Up @@ -161,8 +161,6 @@
JobType.EVALUATION: {
"default": causal_eval_template,
"hf://facebook/bart-large-cnn": bart_eval_template,
"hf://mikeadimech/longformer-qmsum-meeting-summarization": seq2seq_eval_template,
"hf://mrm8488/t5-base-finetuned-summarize-news": seq2seq_eval_template,
"hf://Falconsai/text_summarization": seq2seq_eval_template,
"hf://mistralai/Mistral-7B-Instruct-v0.3": causal_eval_template,
"oai://gpt-4o-mini": oai_eval_template,
Expand Down
28 changes: 0 additions & 28 deletions lumigator/python/mzai/backend/backend/models.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -15,34 +15,6 @@
no_repeat_ngram_size: 3
num_beams: 4

- name: mikeadimech/longformer-qmsum-meeting-summarization
uri: hf://mikeadimech/longformer-qmsum-meeting-summarization
website_url: https://huggingface.co/mikeadimech/longformer-qmsum-meeting-summarization
description: Longformer is a transformer model that is capable of processing long sequences.
info:
parameter_count: 162M
tensor_type: F32
model_size: 648MB
tasks:
- summarization:

- name: mrm8488/t5-base-finetuned-summarize-news
uri: hf://mrm8488/t5-base-finetuned-summarize-news
website_url: https://huggingface.co/mrm8488/t5-base-finetuned-summarize-news
description: Google's T5 base fine-tuned on News Summary dataset for summarization downstream task.
info:
parameter_count: 223M
tensor_type: F32
model_size: 892MB
tasks:
- summarization:
max_length: 200
min_length: 30
length_penalty: 2.0
early_stopping: true
no_repeat_ngram_size: 3
num_beams: 4

- name: Falconsai/text_summarization
uri: hf://Falconsai/text_summarization
website_url: https://huggingface.co/Falconsai/text_summarization
Expand Down
39 changes: 1 addition & 38 deletions lumigator/python/mzai/backend/backend/tests/data/models.json
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
{
"total": 9,
"total": 7,
"items": [
{
"name": "facebook/bart-large-cnn",
Expand All @@ -23,43 +23,6 @@
}
]
},
{
"name": "mikeadimech/longformer-qmsum-meeting-summarization",
"uri": "hf://mikeadimech/longformer-qmsum-meeting-summarization",
"description": "Longformer is a transformer model that is capable of processing long sequences.",
"info": {
"parameter_count": "162M",
"tensor_type": "F32",
"model_size": "648MB"
},
"tasks": [
{
"summarization": null
}
]
},
{
"name": "mrm8488/t5-base-finetuned-summarize-news",
"uri": "hf://mrm8488/t5-base-finetuned-summarize-news",
"description": "Google's T5 base fine-tuned on News Summary dataset for summarization downstream task.",
"info": {
"parameter_count": "223M",
"tensor_type": "F32",
"model_size": "892MB"
},
"tasks": [
{
"summarization": {
"max_length": 200,
"min_length": 30,
"length_penalty": 2,
"early_stopping": true,
"no_repeat_ngram_size": 3,
"num_beams": 4
}
}
]
},
{
"name": "Falconsai/text_summarization",
"uri": "hf://Falconsai/text_summarization",
Expand Down
41 changes: 1 addition & 40 deletions lumigator/python/mzai/sdk/tests/data/models.json
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
{
"total": 9,
"total": 7,
"items": [
{
"name": "facebook/bart-large-cnn",
Expand All @@ -24,45 +24,6 @@
}
]
},
{
"name": "mikeadimech/longformer-qmsum-meeting-summarization",
"uri": "hf://mikeadimech/longformer-qmsum-meeting-summarization",
"website_url": "https://huggingface.co/mikeadimech/longformer-qmsum-meeting-summarization/discussions",
"description": "Longformer is a transformer model that is capable of processing long sequences.",
"info": {
"parameter_count": "162M",
"tensor_type": "F32",
"model_size": "648MB"
},
"tasks": [
{
"summarization": null
}
]
},
{
"name": "mrm8488/t5-base-finetuned-summarize-news",
"uri": "hf://mrm8488/t5-base-finetuned-summarize-news",
"website_url": "https://huggingface.co/mrm8488/t5-base-finetuned-summarize-news",
"description": "Google's T5 base fine-tuned on News Summary dataset for summarization downstream task.",
"info": {
"parameter_count": "223M",
"tensor_type": "F32",
"model_size": "892MB"
},
"tasks": [
{
"summarization": {
"max_length": 200,
"min_length": 30,
"length_penalty": 2,
"early_stopping": true,
"no_repeat_ngram_size": 3,
"num_beams": 4
}
}
]
},
{
"name": "Falconsai/text_summarization",
"uri": "hf://Falconsai/text_summarization",
Expand Down
4 changes: 1 addition & 3 deletions notebooks/assets/model_info.csv
Original file line number Diff line number Diff line change
@@ -1,7 +1,5 @@
model_name,RAM_MiB,RAM_GB
hf://facebook/bart-large-cnn,2709MiB,2.71
hf://mikeadimech/longformer-qmsum-meeting-summarization,2027MiB,2.03
hf://mrm8488/t5-base-finetuned-summarize-news,3085MiB,3.09
hf://Falconsai/text_summarization,1423MiB,1.43
hf://mistralai/Mistral-7B-Instruct-v0.3,30645MiB,30.65
mistral://open-mistral-7b,30645MiB,30.65
Expand All @@ -10,4 +8,4 @@ hf://meta-llama/Meta-Llama-3-8B,34189MiB,34.19
hf://microsoft/Phi-3-mini-4k-instruct,19455MiB,19.46
oai://gpt-4o-mini,,
oai://gpt-4-turbo,,
oai://gpt-3.5-turbo-0125,,
oai://gpt-3.5-turbo-0125,,
Loading

0 comments on commit 74497f6

Please sign in to comment.