language: | |
- hu | |
pipeline_tag: summarization | |
inference: | |
parameters: | |
num_beams: 5 | |
length_penalty: 2 | |
max_length: 128 | |
encoder_no_repeat_ngram_size: 4 | |
no_repeat_ngram_size: 3 | |
datasets: | |
- SZTAKI-HLT/HunSum-1 | |
metrics: | |
- rouge | |
license: apache-2.0 | |
# Model Card for mT5-base-HunSum-1 | |
The mT5-base-HunSum-1 is a Hungarian abstractive summarization model, which was trained on the [SZTAKI-HLT/HunSum-1 dataset](https://huggingface.co/datasets/SZTAKI-HLT/HunSum-1). | |
The model is based on [google/mt5-base](https://huggingface.co/google/mt5-base). | |
## Intended uses & limitations | |
- **Model type:** Text Summarization | |
- **Language(s) (NLP):** Hungarian | |
- **Resource(s) for more information:** | |
- [GitHub Repo](https://github.com/dorinapetra/summarization) | |
## Parameters | |
- **Batch Size:** 12 | |
- **Learning Rate:** 5e-5 | |
- **Weight Decay:** 0.01 | |
- **Warmup Steps:** 3000 | |
- **Epochs:** 10 | |
- **no_repeat_ngram_size:** 3 | |
- **num_beams:** 5 | |
- **early_stopping:** False | |
- **encoder_no_repeat_ngram_size:** 4 | |
## Results | |
| Metric | Value | | |
| :------------ | :------------------------------------------ | | |
| ROUGE-1 | 37.70 | | |
| ROUGE-2 | 11.22 | | |
| ROUGE-L | 24.37 | |