File size: 765 Bytes
3c4b60e 90e0a86 3c4b60e 90e0a86 4d4369f 90e0a86 9020620 90e0a86 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
---
language:
- "ur"
license: "mit"
datasets:
- "Urdu-news-dataset"
---
# GPT-2
Fine tune gpt2 model on Urdu news dataset using a causal language modeling (CLM) objective.
### How to use
You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we
set a seed for reproducibility:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("Imran1/gpt2-urdu-news")
model = AutoModelForCausalLM.from_pretrained("Imran1/gpt2-urdu-news")
```
## Training data
I fine tune gpt2 for downstream task like text generation, only for 1000 sample so it may not be good. Due to resources limitation.
## Evaluation results
training loss 3.042
|