Training in progress, step 250
Browse files- pytorch_model.bin +1 -1
- run.sh +5 -5
- run_speech_recognition_ctc_bnb.py +21 -21
- runs/Jan28_13-50-07_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/events.out.tfevents.1643378396.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.124745.0 +2 -2
- runs/Jan28_19-16-02_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/1643397461.831888/events.out.tfevents.1643397461.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.261029.1 +3 -0
- runs/Jan28_19-16-02_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/events.out.tfevents.1643397461.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.261029.0 +3 -0
- special_tokens_map.json +1 -1
- training_args.bin +1 -1
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1262104049
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c56d8171c7e99a5c2a9e54c0c4428bdc8901de1a9fc1f2248238850b2cae028f
|
3 |
size 1262104049
|
run.sh
CHANGED
@@ -8,17 +8,17 @@ python run_speech_recognition_ctc_bnb.py \
|
|
8 |
--output_dir="./" \
|
9 |
--overwrite_output_dir \
|
10 |
--hub_model_id="NbAiLab/wav2vec2-xlsr-300m-norwegian2" \
|
11 |
-
--num_train_epochs="
|
12 |
-
--per_device_train_batch_size="
|
13 |
-
--per_device_eval_batch_size="
|
14 |
--gradient_accumulation_steps="4" \
|
15 |
--learning_rate="7.5e-5" \
|
16 |
--warmup_steps="2000" \
|
17 |
--evaluation_strategy="steps" \
|
18 |
--length_column_name="input_length" \
|
19 |
--chars_to_ignore , ? . ! \- \; \: \" " % ‘ " � — ’ … – \
|
20 |
-
--save_steps="
|
21 |
-
--eval_steps="
|
22 |
--logging_steps="100" \
|
23 |
--layerdrop="0.0" \
|
24 |
--activation_dropout="0.1" \
|
|
|
8 |
--output_dir="./" \
|
9 |
--overwrite_output_dir \
|
10 |
--hub_model_id="NbAiLab/wav2vec2-xlsr-300m-norwegian2" \
|
11 |
+
--num_train_epochs="100" \
|
12 |
+
--per_device_train_batch_size="8" \
|
13 |
+
--per_device_eval_batch_size="8" \
|
14 |
--gradient_accumulation_steps="4" \
|
15 |
--learning_rate="7.5e-5" \
|
16 |
--warmup_steps="2000" \
|
17 |
--evaluation_strategy="steps" \
|
18 |
--length_column_name="input_length" \
|
19 |
--chars_to_ignore , ? . ! \- \; \: \" " % ‘ " � — ’ … – \
|
20 |
+
--save_steps="250" \
|
21 |
+
--eval_steps="250" \
|
22 |
--logging_steps="100" \
|
23 |
--layerdrop="0.0" \
|
24 |
--activation_dropout="0.1" \
|
run_speech_recognition_ctc_bnb.py
CHANGED
@@ -686,25 +686,25 @@ def main():
|
|
686 |
# Instantiate custom data collator
|
687 |
data_collator = DataCollatorCTCWithPadding(processor=processor)
|
688 |
|
689 |
-
decay_parameters = get_parameter_names(model, [torch.nn.LayerNorm])
|
690 |
-
decay_parameters = [name for name in decay_parameters if "bias" not in name]
|
691 |
-
optimizer_grouped_parameters = [
|
692 |
-
|
693 |
-
|
694 |
-
|
695 |
-
|
696 |
-
|
697 |
-
|
698 |
-
|
699 |
-
|
700 |
-
]
|
701 |
-
optimizer = bnb.optim.Adam8bit(
|
702 |
-
|
703 |
-
|
704 |
-
|
705 |
-
)
|
706 |
-
|
707 |
-
optimizers = (optimizer, None)
|
708 |
|
709 |
# Initialize Trainer
|
710 |
trainer = Trainer(
|
@@ -715,8 +715,8 @@ def main():
|
|
715 |
train_dataset=vectorized_datasets["train"] if training_args.do_train else None,
|
716 |
eval_dataset=vectorized_datasets["eval"] if training_args.do_eval else None,
|
717 |
tokenizer=feature_extractor,
|
718 |
-
optimizers=optimizers,
|
719 |
-
|
720 |
|
721 |
# 8. Finally, we can start training
|
722 |
|
|
|
686 |
# Instantiate custom data collator
|
687 |
data_collator = DataCollatorCTCWithPadding(processor=processor)
|
688 |
|
689 |
+
#decay_parameters = get_parameter_names(model, [torch.nn.LayerNorm])
|
690 |
+
#decay_parameters = [name for name in decay_parameters if "bias" not in name]
|
691 |
+
#optimizer_grouped_parameters = [
|
692 |
+
# {
|
693 |
+
# "params": [p for n, p in model.named_parameters() if n in decay_parameters],
|
694 |
+
# "weight_decay": training_args.weight_decay,
|
695 |
+
# },
|
696 |
+
# {
|
697 |
+
# "params": [p for n, p in model.named_parameters() if n not in decay_parameters],
|
698 |
+
# "weight_decay": 0.0,
|
699 |
+
# },
|
700 |
+
#]
|
701 |
+
#optimizer = bnb.optim.Adam8bit(
|
702 |
+
# params=optimizer_grouped_parameters,
|
703 |
+
# betas=(training_args.adam_beta1, training_args.adam_beta2),
|
704 |
+
# eps=training_args.adam_epsilon,
|
705 |
+
#)
|
706 |
+
|
707 |
+
#optimizers = (optimizer, None)
|
708 |
|
709 |
# Initialize Trainer
|
710 |
trainer = Trainer(
|
|
|
715 |
train_dataset=vectorized_datasets["train"] if training_args.do_train else None,
|
716 |
eval_dataset=vectorized_datasets["eval"] if training_args.do_eval else None,
|
717 |
tokenizer=feature_extractor,
|
718 |
+
#optimizers=optimizers,
|
719 |
+
)
|
720 |
|
721 |
# 8. Finally, we can start training
|
722 |
|
runs/Jan28_13-50-07_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/events.out.tfevents.1643378396.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.124745.0
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2abb1d86b00c8a06e558229507420c951af09afb063ee92e8eb4f72c8b2a4d9c
|
3 |
+
size 18010
|
runs/Jan28_19-16-02_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/1643397461.831888/events.out.tfevents.1643397461.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.261029.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22295f46dbc159d371cb3f6f7811ecb0a43041a21bd080154de37696780d1096
|
3 |
+
size 4786
|
runs/Jan28_19-16-02_job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba/events.out.tfevents.1643397461.job-c93f32d8-97c5-48e7-b5ec-c6c950f627ba.261029.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9193521f5ce2e3c7e6b144d8d2e145fe762c0966a69c1e47edb42c648ad67bc5
|
3 |
+
size 5347
|
special_tokens_map.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
|
|
1 |
+
{"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3055
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:249302cc973fa88a60488bebc95f0dde19627b173724b2e0910323ff0a12fe01
|
3 |
size 3055
|