Model save
Browse files- README.md +2 -6
- adapter_config.json +2 -2
- adapter_model.safetensors +1 -1
- runs/Apr14_22-29-34_ubuntu/events.out.tfevents.1713108584.ubuntu +3 -0
- runs/Apr14_22-32-20_ubuntu/events.out.tfevents.1713108755.ubuntu +3 -0
- runs/Apr15_10-46-36_ubuntu/events.out.tfevents.1713152828.ubuntu +3 -0
- runs/Apr15_11-06-02_ubuntu/events.out.tfevents.1713153964.ubuntu +3 -0
- tokenizer_config.json +1 -0
- training_args.bin +1 -1
README.md
CHANGED
@@ -36,20 +36,16 @@ More information needed
|
|
36 |
### Training hyperparameters
|
37 |
|
38 |
The following hyperparameters were used during training:
|
39 |
-
- learning_rate: 0.
|
40 |
- train_batch_size: 8
|
41 |
- eval_batch_size: 8
|
42 |
-
- seed:
|
43 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
44 |
- lr_scheduler_type: linear
|
45 |
- lr_scheduler_warmup_steps: 1
|
46 |
- num_epochs: 1
|
47 |
- mixed_precision_training: Native AMP
|
48 |
|
49 |
-
### Training results
|
50 |
-
|
51 |
-
|
52 |
-
|
53 |
### Framework versions
|
54 |
|
55 |
- PEFT 0.10.0
|
|
|
36 |
### Training hyperparameters
|
37 |
|
38 |
The following hyperparameters were used during training:
|
39 |
+
- learning_rate: 0.0002
|
40 |
- train_batch_size: 8
|
41 |
- eval_batch_size: 8
|
42 |
+
- seed: 4269
|
43 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
44 |
- lr_scheduler_type: linear
|
45 |
- lr_scheduler_warmup_steps: 1
|
46 |
- num_epochs: 1
|
47 |
- mixed_precision_training: Native AMP
|
48 |
|
|
|
|
|
|
|
|
|
49 |
### Framework versions
|
50 |
|
51 |
- PEFT 0.10.0
|
adapter_config.json
CHANGED
@@ -20,8 +20,8 @@
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
-
"
|
24 |
-
"
|
25 |
],
|
26 |
"task_type": "CAUSAL_LM",
|
27 |
"use_dora": false,
|
|
|
20 |
"rank_pattern": {},
|
21 |
"revision": null,
|
22 |
"target_modules": [
|
23 |
+
"q_proj",
|
24 |
+
"v_proj"
|
25 |
],
|
26 |
"task_type": "CAUSAL_LM",
|
27 |
"use_dora": false,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 109069176
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79cf4849132a2098f10c59f47222c8f12ad2fdd76c6f3bd311012f57446bf5bf
|
3 |
size 109069176
|
runs/Apr14_22-29-34_ubuntu/events.out.tfevents.1713108584.ubuntu
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53a1884652aae41ca380cd25fd44d348bd14320ae9a90c645e902e6adcdfd202
|
3 |
+
size 5053
|
runs/Apr14_22-32-20_ubuntu/events.out.tfevents.1713108755.ubuntu
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41f90e34e1ad0b245c85cbe4d0ef905f515b9534e56d52bd49a89baafbf1aa49
|
3 |
+
size 83032
|
runs/Apr15_10-46-36_ubuntu/events.out.tfevents.1713152828.ubuntu
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b2d2721602df64e0525a4b4b1e22128e6a00cf572ba588493922af04c103f86a
|
3 |
+
size 10514
|
runs/Apr15_11-06-02_ubuntu/events.out.tfevents.1713153964.ubuntu
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:00fafeaf0b125ba9692a139c0872a9e378c42f03ce425160250307ffbad9339d
|
3 |
+
size 591912
|
tokenizer_config.json
CHANGED
@@ -32,6 +32,7 @@
|
|
32 |
"clean_up_tokenization_spaces": false,
|
33 |
"eos_token": "</s>",
|
34 |
"legacy": true,
|
|
|
35 |
"model_max_length": 1000000000000000019884624838656,
|
36 |
"pad_token": "</s>",
|
37 |
"sp_model_kwargs": {},
|
|
|
32 |
"clean_up_tokenization_spaces": false,
|
33 |
"eos_token": "</s>",
|
34 |
"legacy": true,
|
35 |
+
"max_seq_length": 1024,
|
36 |
"model_max_length": 1000000000000000019884624838656,
|
37 |
"pad_token": "</s>",
|
38 |
"sp_model_kwargs": {},
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4984
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7ef14cbbaa82c9476964a862acb20d4f9512f4fdd2dd3044e9cf6352408a41d8
|
3 |
size 4984
|