new attempt
Browse files
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "./",
|
3 |
"architectures": [
|
4 |
"XLMRobertaForMaskedLM"
|
5 |
],
|
@@ -21,7 +21,7 @@
|
|
21 |
"pad_token_id": 1,
|
22 |
"position_embedding_type": "absolute",
|
23 |
"torch_dtype": "float32",
|
24 |
-
"transformers_version": "4.
|
25 |
"type_vocab_size": 1,
|
26 |
"use_cache": true,
|
27 |
"vocab_size": 250002
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "./config.json",
|
3 |
"architectures": [
|
4 |
"XLMRobertaForMaskedLM"
|
5 |
],
|
|
|
21 |
"pad_token_id": 1,
|
22 |
"position_embedding_type": "absolute",
|
23 |
"torch_dtype": "float32",
|
24 |
+
"transformers_version": "4.24.0",
|
25 |
"type_vocab_size": 1,
|
26 |
"use_cache": true,
|
27 |
"vocab_size": 250002
|
events.out.tfevents.1672819976.t1v-n-0853dee6-w-3.291836.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4189a298dc3ab9be3250fd74db70fac076b8e11d308c2afba23e19d047a123a8
|
3 |
+
size 22598
|
events.out.tfevents.1672820882.t1v-n-0853dee6-w-3.322000.0.v2
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9f7294f94d22c80b866f0559f6606ac68f6f6f386918848c1f7f4ba68e161340
|
3 |
+
size 7637
|
flax_model.msgpack
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8a04d8032d9ff18b1b727c98ee3bea82de1d908107d257d94486dba8650f5680
|
3 |
+
size 1113187999
|
run.sh
CHANGED
@@ -1,6 +1,5 @@
|
|
1 |
|
2 |
python run_mlm_flax_stream.py \
|
3 |
-
--output_dir="./" \
|
4 |
--hub_model_id="NbAiLab/nb-roberta-base-scandi" \
|
5 |
--model_name_or_path="xlm-roberta-base" \
|
6 |
--config_name="./config.json" \
|
@@ -12,6 +11,7 @@ python run_mlm_flax_stream.py \
|
|
12 |
--per_device_eval_batch_size="16" \
|
13 |
--learning_rate="4e-4" \
|
14 |
--warmup_steps="1000" \
|
|
|
15 |
--overwrite_output_dir \
|
16 |
--num_train_steps="10000" \
|
17 |
--adam_beta1="0.9" \
|
@@ -19,5 +19,6 @@ python run_mlm_flax_stream.py \
|
|
19 |
--logging_steps="50" \
|
20 |
--save_steps="50" \
|
21 |
--eval_steps="50" \
|
|
|
22 |
--dtype="bfloat16" \
|
23 |
--push_to_hub
|
|
|
1 |
|
2 |
python run_mlm_flax_stream.py \
|
|
|
3 |
--hub_model_id="NbAiLab/nb-roberta-base-scandi" \
|
4 |
--model_name_or_path="xlm-roberta-base" \
|
5 |
--config_name="./config.json" \
|
|
|
11 |
--per_device_eval_batch_size="16" \
|
12 |
--learning_rate="4e-4" \
|
13 |
--warmup_steps="1000" \
|
14 |
+
--use_auth_token \
|
15 |
--overwrite_output_dir \
|
16 |
--num_train_steps="10000" \
|
17 |
--adam_beta1="0.9" \
|
|
|
19 |
--logging_steps="50" \
|
20 |
--save_steps="50" \
|
21 |
--eval_steps="50" \
|
22 |
+
push_to_--output_dir="./" \
|
23 |
--dtype="bfloat16" \
|
24 |
--push_to_hub
|
run_mlm_flax_stream.py
CHANGED
@@ -655,13 +655,18 @@ if __name__ == "__main__":
|
|
655 |
)
|
656 |
print(f"Saving weights and logs of step {step+1}. \nThe result is saved to {training_args.output_folder} by worker {jax.process_index()}.")
|
657 |
except:
|
658 |
-
|
659 |
model.save_pretrained(
|
660 |
training_args.output_dir,
|
661 |
params=params
|
662 |
)
|
663 |
print("Problems pushing this to the hub. The bug should be fixed.")
|
664 |
-
|
|
|
|
|
|
|
|
|
|
|
665 |
# update tqdm bar
|
666 |
steps.update(1)
|
667 |
|
|
|
655 |
)
|
656 |
print(f"Saving weights and logs of step {step+1}. \nThe result is saved to {training_args.output_folder} by worker {jax.process_index()}.")
|
657 |
except:
|
658 |
+
breakpoint()
|
659 |
model.save_pretrained(
|
660 |
training_args.output_dir,
|
661 |
params=params
|
662 |
)
|
663 |
print("Problems pushing this to the hub. The bug should be fixed.")
|
664 |
+
else:
|
665 |
+
model.save_pretrained(
|
666 |
+
training_args.output_dir,
|
667 |
+
params=params
|
668 |
+
)
|
669 |
+
|
670 |
# update tqdm bar
|
671 |
steps.update(1)
|
672 |
|