fozziethebeat commited on
Commit
40e02d7
·
verified ·
1 Parent(s): c1f182c

End of training

Browse files
Files changed (1) hide show
  1. README.md +181 -0
README.md ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: meta-llama/Meta-Llama-3-8B-Instruct
3
+ library_name: peft
4
+ license: llama3
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ model-index:
9
+ - name: l3bgi-sft-qlora-r64
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
17
+ <details><summary>See axolotl config</summary>
18
+
19
+ axolotl version: `0.4.1`
20
+ ```yaml
21
+ # Allow cli options to override these settings.
22
+ strict: false
23
+
24
+ # Base model settings.
25
+ base_model: meta-llama/Meta-Llama-3-8B-Instruct
26
+ tokenizer_config: meta-llama/Meta-Llama-3-8B-Instruct
27
+ model_type: AutoModelForCausalLM
28
+
29
+ # Wandb settings
30
+ wandb_entity: collinear
31
+ wandb_project: template-training
32
+ wandb_name: l3smi-sft-qlora-r64
33
+
34
+ # Output settings
35
+ save_safetensors: true
36
+ hub_model_id: fozziethebeat/l3bgi-sft-qlora-r64
37
+ dataset_prepared_path: data/l3bgi-sft-qlora-r64
38
+ output_dir: models/l3bgi-sft-qlora-r64
39
+
40
+ # Data format settings
41
+ chat_template: llama3
42
+ datasets:
43
+ - path: fozziethebeat/alpaca_messages_2k_test
44
+ split: train
45
+ type: chat_template
46
+ chat_template: llama3
47
+ field_messages: messages
48
+ message_field_role: role
49
+ message_field_content: content
50
+ test_datasets:
51
+ - path: fozziethebeat/alpaca_messages_2k_test
52
+ split: test
53
+ type: chat_template
54
+ chat_template: llama3
55
+ field_messages: messages
56
+ message_field_role: role
57
+ message_field_content: content
58
+
59
+ # Data packing settings
60
+ sequence_len: 512
61
+ train_on_inputs: false
62
+ pad_to_sequence_len: true
63
+ group_by_length: false
64
+ sample_packing: false
65
+ eval_sample_packing: false
66
+
67
+ # Adapter settings
68
+ adapter: qlora
69
+ lora_model_dir:
70
+ load_in_8bit: false
71
+ load_in_4bit: true
72
+ lora_r: 64
73
+ lora_alpha: 16
74
+ lora_dropout: 0.05
75
+ lora_target_linear: true
76
+ lora_fan_in_fan_out:
77
+ lora_target_modules:
78
+ - gate_proj
79
+ - down_proj
80
+ - up_proj
81
+ - q_proj
82
+ - v_proj
83
+ - k_proj
84
+ - o_proj
85
+
86
+ # Computation Format settings
87
+ bf16: true
88
+ fp16:
89
+ tf32: false
90
+
91
+ # Trainer settings
92
+ gradient_accumulation_steps: 2
93
+ micro_batch_size: 2
94
+ num_epochs: 2
95
+ optimizer: adamw_torch
96
+ lr_scheduler: cosine
97
+ learning_rate: 1e-5
98
+ loss_watchdog_threshold: 5.0
99
+ loss_watchdog_patience: 3
100
+
101
+ gradient_checkpointing: true
102
+ early_stopping_patience:
103
+ resume_from_checkpoint:
104
+ local_rank:
105
+ logging_steps: 1
106
+ xformers_attention:
107
+ #flash_attention: true
108
+
109
+ warmup_steps: 10
110
+ eval_table_size:
111
+ eval_max_new_tokens: 128
112
+ evals_per_epoch: 4
113
+ saves_per_epoch: 1
114
+ debug:
115
+ weight_decay: 0.01
116
+ special_tokens:
117
+ pad_token: <|end_of_text|>
118
+ deepspeed:
119
+ fsdp:
120
+
121
+ ```
122
+
123
+ </details><br>
124
+
125
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/collinear/template-training/runs/pav37wt6)
126
+ # l3bgi-sft-qlora-r64
127
+
128
+ This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on the None dataset.
129
+ It achieves the following results on the evaluation set:
130
+ - Loss: 1.0220
131
+
132
+ ## Model description
133
+
134
+ More information needed
135
+
136
+ ## Intended uses & limitations
137
+
138
+ More information needed
139
+
140
+ ## Training and evaluation data
141
+
142
+ More information needed
143
+
144
+ ## Training procedure
145
+
146
+ ### Training hyperparameters
147
+
148
+ The following hyperparameters were used during training:
149
+ - learning_rate: 1e-05
150
+ - train_batch_size: 2
151
+ - eval_batch_size: 2
152
+ - seed: 42
153
+ - gradient_accumulation_steps: 2
154
+ - total_train_batch_size: 4
155
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
156
+ - lr_scheduler_type: cosine
157
+ - lr_scheduler_warmup_steps: 10
158
+ - num_epochs: 2
159
+
160
+ ### Training results
161
+
162
+ | Training Loss | Epoch | Step | Validation Loss |
163
+ |:-------------:|:------:|:----:|:---------------:|
164
+ | 1.0859 | 0.0022 | 1 | 1.3374 |
165
+ | 0.9847 | 0.2497 | 111 | 1.1122 |
166
+ | 1.203 | 0.4994 | 222 | 1.0451 |
167
+ | 1.3916 | 0.7492 | 333 | 1.0307 |
168
+ | 0.7893 | 0.9989 | 444 | 1.0251 |
169
+ | 1.0244 | 1.2486 | 555 | 1.0228 |
170
+ | 0.6814 | 1.4983 | 666 | 1.0221 |
171
+ | 0.9408 | 1.7480 | 777 | 1.0224 |
172
+ | 1.0832 | 1.9978 | 888 | 1.0220 |
173
+
174
+
175
+ ### Framework versions
176
+
177
+ - PEFT 0.11.1
178
+ - Transformers 4.43.0.dev0
179
+ - Pytorch 2.3.1+cu121
180
+ - Datasets 2.19.1
181
+ - Tokenizers 0.19.1