File size: 765 Bytes
3c4b60e
90e0a86
 
 
 
3c4b60e
90e0a86
 
 
 
 
 
4d4369f
90e0a86
 
 
 
 
 
 
 
 
 
 
 
 
 
9020620
90e0a86
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
---

language: 
    - "ur"
license: "mit"
datasets:
   - "Urdu-news-dataset"
   
---


# GPT-2
Fine tune gpt2 model on Urdu news dataset using a causal language modeling (CLM) objective.

### How to use

You can use this model directly with a pipeline for text generation. Since the generation relies on some randomness, we
set a seed for reproducibility:

```python
from transformers import AutoTokenizer, AutoModelForCausalLM

tokenizer = AutoTokenizer.from_pretrained("Imran1/gpt2-urdu-news")

model = AutoModelForCausalLM.from_pretrained("Imran1/gpt2-urdu-news")
```
## Training data
I fine tune gpt2 for downstream task like text generation, only for 1000 sample so it may not be good. Due to resources limitation.


## Evaluation results
training loss 3.042