sakren commited on
Commit
be78909
1 Parent(s): 48b2b40

sakren/distil-bert-imeocap

Browse files
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: distilbert-base-uncased
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - f1
8
+ - precision
9
+ - recall
10
+ - accuracy
11
+ model-index:
12
+ - name: distil-bert-imeocap
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # distil-bert-imeocap
20
+
21
+ This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 1.8186
24
+ - F1: 0.6341
25
+ - Precision: 0.6365
26
+ - Recall: 0.6365
27
+ - Accuracy: 0.6365
28
+
29
+ ## Model description
30
+
31
+ More information needed
32
+
33
+ ## Intended uses & limitations
34
+
35
+ More information needed
36
+
37
+ ## Training and evaluation data
38
+
39
+ More information needed
40
+
41
+ ## Training procedure
42
+
43
+ ### Training hyperparameters
44
+
45
+ The following hyperparameters were used during training:
46
+ - learning_rate: 2e-05
47
+ - train_batch_size: 64
48
+ - eval_batch_size: 64
49
+ - seed: 42
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - num_epochs: 15
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | F1 | Precision | Recall | Accuracy |
57
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:---------:|:------:|:--------:|
58
+ | 0.1961 | 1.0 | 74 | 1.6080 | 0.6314 | 0.6285 | 0.6385 | 0.6385 |
59
+ | 0.1845 | 2.0 | 148 | 1.7125 | 0.6298 | 0.6317 | 0.6385 | 0.6385 |
60
+ | 0.1717 | 3.0 | 222 | 1.9402 | 0.6226 | 0.6364 | 0.6385 | 0.6385 |
61
+ | 0.176 | 4.0 | 296 | 1.8028 | 0.6169 | 0.6253 | 0.6192 | 0.6192 |
62
+ | 0.1679 | 5.0 | 370 | 1.6948 | 0.6243 | 0.6285 | 0.625 | 0.625 |
63
+ | 0.168 | 6.0 | 444 | 1.8304 | 0.6317 | 0.6336 | 0.6385 | 0.6385 |
64
+ | 0.1617 | 7.0 | 518 | 1.7457 | 0.6286 | 0.6310 | 0.6308 | 0.6308 |
65
+ | 0.1677 | 8.0 | 592 | 1.8071 | 0.6422 | 0.6382 | 0.65 | 0.65 |
66
+ | 0.171 | 9.0 | 666 | 1.8177 | 0.6323 | 0.6326 | 0.6385 | 0.6385 |
67
+ | 0.1683 | 10.0 | 740 | 1.8265 | 0.6347 | 0.6370 | 0.6365 | 0.6365 |
68
+ | 0.1808 | 11.0 | 814 | 1.7734 | 0.6304 | 0.6365 | 0.6308 | 0.6308 |
69
+ | 0.1757 | 12.0 | 888 | 1.7727 | 0.6244 | 0.6296 | 0.6231 | 0.6231 |
70
+ | 0.1897 | 13.0 | 962 | 1.8449 | 0.6374 | 0.6377 | 0.6404 | 0.6404 |
71
+ | 0.1674 | 14.0 | 1036 | 1.8244 | 0.6455 | 0.6462 | 0.6481 | 0.6481 |
72
+ | 0.1746 | 15.0 | 1110 | 1.8186 | 0.6341 | 0.6365 | 0.6365 | 0.6365 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.39.3
78
+ - Pytorch 2.1.2
79
+ - Datasets 2.18.0
80
+ - Tokenizers 0.15.2
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3a8f31b11436fd57312dd28d878e42bd806354e74488fb1bf1e8d9366d770cc
3
  size 267844872
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2a8c90016ee90844ab468a5e6c6f4cdf9daea495baa3482063a2ceb04bbe0a50
3
  size 267844872
runs/May15_17-13-56_95a092faa389/events.out.tfevents.1715793392.95a092faa389.35.5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a2f56fdcae321095670afabfa6c08607d3b030d916567380ec4ae7fea6356c93
3
- size 13723
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6a016d3f9b6fa412b34096c2795b884d73a1697d461a43c6c8181336c2bee75
3
+ size 15443
runs/May15_17-13-56_95a092faa389/events.out.tfevents.1715793751.95a092faa389.35.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae5aa8154de4395bb3b89b7e78b6666b79bed9d181c02093baaf0a001a63582b
3
+ size 560