--- datasets: - jaehy12/news3 language: - ko base_model: - ibm-granite/granite-3.1-2b-instruct library_name: transformers --- ## Model Details **Granite-3.1-2B-instruct-KR-Summarization** Granite-3.1-2B-instruct-KR-Summarization is continued pretrained(fully fine-tuned) language model based on Granite-3.1-2B-Instruct. This model is trained fully with publicily available resource at HuggingFace dataset hub, preprocessed Korean texts. The train was done on A6000 48GB * 4. **Model developers** Dongwook Min (mindw96) **Variations** Granite-3.1-2B-instruct-KR-Summarization comes in one size — 2B. **Input** Models input text only. **Output** Models generate text only. **Model Architecture** Granite 3.1 is an auto-regressive language model that uses an optimized transformer architecture. **Model Release Date** 02.01.2025. **Capabilities** * Summarization