File size: 1,996 Bytes
7d54bbe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
46ab38f
 
 
7d54bbe
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e5154a3
2837859
 
 
 
e6f7c80
e824e61
21e0ff5
 
 
 
e2a7611
a17711e
eba5663
64b457d
8320b0f
46ab38f
7d54bbe
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
---
license: apache-2.0
base_model: distilgpt2
tags:
- generated_from_keras_callback
model-index:
- name: EngTig/distilgpt2-finetuned-wikitext2
  results: []
---

<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->

# EngTig/distilgpt2-finetuned-wikitext2

This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 2.3661
- Validation Loss: 4.0640
- Epoch: 16

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- optimizer: {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}
- training_precision: float32

### Training results

| Train Loss | Validation Loss | Epoch |
|:----------:|:---------------:|:-----:|
| 2.9937     | 3.8775          | 0     |
| 2.9426     | 3.8763          | 1     |
| 2.8926     | 3.8593          | 2     |
| 2.8445     | 3.8982          | 3     |
| 2.8090     | 3.9044          | 4     |
| 2.7511     | 3.9337          | 5     |
| 2.7140     | 3.9265          | 6     |
| 2.6655     | 3.9483          | 7     |
| 2.6443     | 3.9490          | 8     |
| 2.6153     | 3.9458          | 9     |
| 2.5699     | 3.9660          | 10    |
| 2.5262     | 3.9897          | 11    |
| 2.5002     | 4.0219          | 12    |
| 2.4636     | 4.0540          | 13    |
| 2.4327     | 4.0224          | 14    |
| 2.3945     | 4.0364          | 15    |
| 2.3661     | 4.0640          | 16    |


### Framework versions

- Transformers 4.38.2
- TensorFlow 2.15.0
- Datasets 2.18.0
- Tokenizers 0.15.2