|
--- |
|
license: apache-2.0 |
|
datasets: |
|
- knkarthick/samsum |
|
language: |
|
- en |
|
metrics: |
|
- rouge |
|
base_model: |
|
- google-t5/t5-base |
|
pipeline_tag: summarization |
|
tags: |
|
- generated_from_text |
|
--- |
|
## Summarizer |
|
This model is a fine-tuned version of google-t5/t5-base on a knkarthick/samsum dataset. It achieves the following results on the evaluation set: |
|
|
|
• ROUGE-1: 51.41 |
|
|
|
• ROUGE-2: 26.72 |
|
|
|
• ROUGE-L: 42.15 |
|
|
|
• ROUGE-Lsum: 42.17 |
|
|
|
## Training hyperparameters |
|
|
|
The following hyperparameters were used during training: |
|
|
|
• learning_rate: 5e-5 |
|
|
|
• train_batch_size: 2 |
|
|
|
• gradient_accumulation_steps=2 |
|
|
|
• seed: 42 |
|
|
|
• weight_decay=0.01 |
|
|
|
• lr_scheduler_type: linear |
|
|
|
• warmup_ratio=0.1 |
|
|
|
• num_epochs: 3 |
|
|
|
## Training results |
|
| Training Loss | Epoch | Step | |
|
| ------------- | ----- | ----- | |
|
| 1.491800 | 1 | 3650 | |
|
| 1.404900 | 2 | 7350 | |
|
| 1.322800 | 3 | 11000 | |
|
|
|
|
|
## Framework versions |
|
• transformers: 4.56.0 |
|
|
|
• pytorch: 2.0.1+cu118 |
|
|
|
• datasets: 2.14.4 |
|
|
|
• tokenizers: 0.13.3 |
|
|