coinplusfire commited on
Commit
0d85c37
·
verified ·
1 Parent(s): a47788f

coinplusfire/coinplusfire_llm

Browse files
README.md CHANGED
@@ -3,7 +3,7 @@ license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - generated_from_trainer
6
- base_model: TheBloke/Mistral-7B-Instruct-v0.2-GPTQ
7
  model-index:
8
  - name: coinplusfire_llm
9
  results: []
@@ -14,9 +14,9 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # coinplusfire_llm
16
 
17
- This model is a fine-tuned version of [TheBloke/Mistral-7B-Instruct-v0.2-GPTQ](https://huggingface.co/TheBloke/Mistral-7B-Instruct-v0.2-GPTQ) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 1.8949
20
 
21
  ## Model description
22
 
@@ -35,32 +35,23 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 0.0002
39
- - train_batch_size: 4
40
- - eval_batch_size: 4
41
  - seed: 42
42
  - gradient_accumulation_steps: 4
43
- - total_train_batch_size: 16
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 2
47
- - num_epochs: 10
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
- | 4.5942 | 0.92 | 3 | 3.9658 |
55
- | 4.0555 | 1.85 | 6 | 3.4516 |
56
- | 3.4878 | 2.77 | 9 | 3.0037 |
57
- | 2.2707 | 4.0 | 13 | 2.5762 |
58
- | 2.6976 | 4.92 | 16 | 2.3493 |
59
- | 2.4044 | 5.85 | 19 | 2.1719 |
60
- | 2.1839 | 6.77 | 22 | 2.0236 |
61
- | 1.5376 | 8.0 | 26 | 1.9475 |
62
- | 2.0032 | 8.92 | 29 | 1.9066 |
63
- | 1.3929 | 9.23 | 30 | 1.8949 |
64
 
65
 
66
  ### Framework versions
 
3
  library_name: peft
4
  tags:
5
  - generated_from_trainer
6
+ base_model: mistralai/Mistral-7B-Instruct-v0.2
7
  model-index:
8
  - name: coinplusfire_llm
9
  results: []
 
14
 
15
  # coinplusfire_llm
16
 
17
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.2](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 2.0778
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - learning_rate: 0.0005
39
+ - train_batch_size: 8
40
+ - eval_batch_size: 8
41
  - seed: 42
42
  - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 32
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 2
47
+ - num_epochs: 1
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:-----:|:----:|:---------------:|
54
+ | 2.891 | 0.99 | 51 | 2.0778 |
 
 
 
 
 
 
 
 
 
55
 
56
 
57
  ### Framework versions
runs/Apr13_08-31-02_22321abdc3e0/events.out.tfevents.1712997062.22321abdc3e0.5594.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31e62ded1f842ae8c7b4776e8e5b94737eacd693c7a9c2fd06bbe61d40c91854
3
+ size 5399
runs/Apr13_08-31-02_22321abdc3e0/events.out.tfevents.1712997386.22321abdc3e0.5594.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:474d2edf91e58b1045fbf17bf753790315ff0f62df4b6b536f16c9a2fd8e9b76
3
+ size 5399
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aa7bb98caab6fee79d139bc154e2b7e78d735c87e2701b4e3711d66235a0b05f
3
  size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae457afd3ec9ae8e7172b590111d1eb73dfe6e235d07cca6d4ee331a3356f179
3
  size 4856