|
--- |
|
license: openrail |
|
language: |
|
- tl |
|
tags: |
|
- langauge |
|
- gpt |
|
- remake |
|
- v2 |
|
- pytorch |
|
- pickle |
|
- gpt2 |
|
- open sourced |
|
pipeline_tag: text-generation |
|
--- |
|
Colab used to train this model ππ [gpt remaker](https://colab.research.google.com/drive/1O9uFQVP9EUhguwhx2qD4pk9PbRCdnijE?usp=sharing) |
|
Both training and inferencing are included in the colab. Happy coding! |
|
|
|
# Model Information |
|
- Model Name: GPTagalog |
|
- Version: 2 |
|
- Training Iterations: 143,000 |
|
- Learning Rate: 6e-4 |
|
- Language: Tagalog |
|
- Compatibility: Pickle (pkl) format (cuda) |
|
- Model Size: 30MB |
|
- Training Time: Approx 2 hours and 30 minutes |
|
- Usage: Experimental, not suitable for commercial purposes |
|
|
|
# Model Description |
|
This was designed to explore the capabilities of training a language model on a small dataset and to see how it performs in generating text in the Tagalog language. |
|
|
|
# Training Details |
|
Iterations and Epochs: GPTagalog was trained for 143,000 iterations over 60 epochs. This extended training period aimed to refine its language generation abilities. |
|
|
|
Learning Rate: The model was trained with a learning rate of 6e-4, which was chosen to optimize learning and convergence. |
|
|
|
Model Size: GPTagalog is relatively small with a file size of 30MB. This small size is due to its experimental nature and limited resources. |
|
|
|
# Usage Guidelines |
|
Experimental Use: GPTagalog Version 2 is an experimental model and is not recommended for commercial purposes. It may have limitations in generating coherent and contextually accurate text. |
|
|
|
Resource Constraints: Due to resource limitations, the model's training was limited to 143,000 iterations and a maximum training time of 6 hours. This is considerably shorter than the training duration of larger models like GPT-2, which has 143 million parameters and takes several days to train. |