harouzie commited on
Commit
4627022
1 Parent(s): 056f98a

harouzie/mt5-small-translation-en2vi

Browse files
Files changed (5) hide show
  1. README.md +7 -7
  2. config.json +1 -1
  3. generation_config.json +1 -1
  4. pytorch_model.bin +1 -1
  5. training_args.bin +1 -1
README.md CHANGED
@@ -17,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 11.7288
21
- - Bleu: 0.0178
22
- - Gen Len: 2.8244
23
 
24
  ## Model description
25
 
@@ -52,12 +52,12 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
54
  |:-------------:|:-----:|:----:|:---------------:|:------:|:-------:|
55
- | No log | 1.0 | 317 | 11.7877 | 0.0797 | 2.8268 |
56
 
57
 
58
  ### Framework versions
59
 
60
- - Transformers 4.33.1
61
- - Pytorch 2.0.1+cu118
62
- - Datasets 2.14.5
63
  - Tokenizers 0.13.3
 
17
 
18
  This model is a fine-tuned version of [google/mt5-small](https://huggingface.co/google/mt5-small) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 22.2680
21
+ - Bleu: 0.0468
22
+ - Gen Len: 2.2283
23
 
24
  ## Model description
25
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
54
  |:-------------:|:-----:|:----:|:---------------:|:------:|:-------:|
55
+ | No log | 0.99 | 63 | 22.7318 | 0.0494 | 2.2717 |
56
 
57
 
58
  ### Framework versions
59
 
60
+ - Transformers 4.33.0
61
+ - Pytorch 2.0.0
62
+ - Datasets 2.1.0
63
  - Tokenizers 0.13.3
config.json CHANGED
@@ -26,7 +26,7 @@
26
  "tie_word_embeddings": false,
27
  "tokenizer_class": "T5Tokenizer",
28
  "torch_dtype": "float32",
29
- "transformers_version": "4.33.1",
30
  "use_cache": true,
31
  "vocab_size": 250102
32
  }
 
26
  "tie_word_embeddings": false,
27
  "tokenizer_class": "T5Tokenizer",
28
  "torch_dtype": "float32",
29
+ "transformers_version": "4.33.0",
30
  "use_cache": true,
31
  "vocab_size": 250102
32
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
- "transformers_version": "4.33.1"
6
  }
 
2
  "decoder_start_token_id": 0,
3
  "eos_token_id": 1,
4
  "pad_token_id": 0,
5
+ "transformers_version": "4.33.0"
6
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:90096790f3c10d551db32e7ae8a04b924f33b9918515a766898c86aeba27135b
3
  size 1200731653
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4139186ba0610ba6c17f97b333f48c5c093b0191cf5738bb58c948686e97928f
3
  size 1200731653
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aabf268e1f54ced93209880d1a5927a76c3ea71ff13e621b27fc39ec5933255d
3
  size 4219
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:837f379329c98ffe2c9e362a6f92fc81dfb8b25d417c9e16b9ecf6703e96d551
3
  size 4219