|
--- |
|
datasets: |
|
- jaehy12/news3 |
|
language: |
|
- ko |
|
base_model: |
|
- ibm-granite/granite-3.1-2b-instruct |
|
library_name: transformers |
|
--- |
|
## Model Details |
|
|
|
**Granite-3.1-2B-instruct-KR-Summarization** |
|
|
|
Granite-3.1-2B-instruct-KR-Summarization is continued pretrained(fully fine-tuned) language model based on Granite-3.1-2B-Instruct. |
|
|
|
This model is trained fully with publicily available resource at HuggingFace dataset hub, preprocessed Korean texts. |
|
|
|
The train was done on A6000 48GB * 4. |
|
|
|
**Model developers** Dongwook Min (mindw96) |
|
|
|
**Variations** Granite-3.1-2B-instruct-KR-Summarization comes in one size — 2B. |
|
|
|
**Input** Models input text only. |
|
|
|
**Output** Models generate text only. |
|
|
|
**Model Architecture** Granite 3.1 is an auto-regressive language model that uses an optimized transformer architecture. |
|
|
|
**Model Release Date** 02.01.2025. |
|
|
|
**Capabilities** |
|
* Summarization |