|
--- |
|
|
|
language: |
|
- "ur" |
|
license: "mit" |
|
datasets: |
|
- "Urdu-news-dataset" |
|
|
|
--- |
|
|
|
|
|
# GPT-2 |
|
Fine tune gpt2 model on Urdu news dataset using a causal language modeling (CLM) objective. |
|
|
|
### How to use |
|
|
|
You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we |
|
set a seed for reproducibility: |
|
|
|
```python |
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("Imran1/gpt2-urdu-news") |
|
|
|
model = AutoModelForCausalLM.from_pretrained("Imran1/gpt2-urdu-news") |
|
``` |
|
## Training data |
|
I fine tune gpt2 for downstream task like text generation, only for 1000 sample so it may not be good. Due to resources limitation. |
|
|
|
|
|
## Evaluation results |
|
training loss 3.042 |
|
|