0Tick commited on
Commit
7897c96
1 Parent(s): 75fc7c6

Trained with tags scrambled

Browse files

This is a new version of the model where the tags of each image were scrambled before training so it should have better completion features.

config.json CHANGED
@@ -40,7 +40,7 @@
40
  }
41
  },
42
  "torch_dtype": "float32",
43
- "transformers_version": "4.27.0.dev0",
44
  "use_cache": true,
45
  "vocab_size": 50257
46
  }
 
40
  }
41
  },
42
  "torch_dtype": "float32",
43
+ "transformers_version": "4.30.0.dev0",
44
  "use_cache": true,
45
  "vocab_size": 50257
46
  }
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ t
3
  Ġ a
4
  h e
 
1
+ #version: 0.2
2
  Ġ t
3
  Ġ a
4
  h e
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4c5230f37a0f613c037dd8acdb4372be3eca15a188cde085e54f42f209e30695
3
- size 333970169
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:794383dd0da7fc82de3d9453c8fa0731a3a9b2bc355fefa67329dd49b4b89db3
3
+ size 327674773
runs/Jun07_18-24-50_86e9400747d2/events.out.tfevents.1686162358.86e9400747d2.1933.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9346d9e5048369d1962f2420f8d29b047e92d85be1e9b713bdd0faee2e16979b
3
+ size 7088
runs/Jun07_18-24-50_86e9400747d2/events.out.tfevents.1686172708.86e9400747d2.1933.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c09a9b1deba394ba794c59b8e8b88c1e5925b183d33c2e1a1ce581a9eac369bb
3
+ size 411
tokenizer.json CHANGED
@@ -39,6 +39,7 @@
39
  "continuing_subword_prefix": "",
40
  "end_of_word_suffix": "",
41
  "fuse_unk": false,
 
42
  "vocab": {
43
  "!": 0,
44
  "\"": 1,
 
39
  "continuing_subword_prefix": "",
40
  "end_of_word_suffix": "",
41
  "fuse_unk": false,
42
+ "byte_fallback": false,
43
  "vocab": {
44
  "!": 0,
45
  "\"": 1,
tokenizer_config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<|endoftext|>",
 
4
  "eos_token": "<|endoftext|>",
5
  "model_max_length": 1024,
6
- "special_tokens_map_file": null,
7
  "tokenizer_class": "GPT2Tokenizer",
8
  "unk_token": "<|endoftext|>"
9
  }
 
1
  {
2
  "add_prefix_space": false,
3
  "bos_token": "<|endoftext|>",
4
+ "clean_up_tokenization_spaces": true,
5
  "eos_token": "<|endoftext|>",
6
  "model_max_length": 1024,
 
7
  "tokenizer_class": "GPT2Tokenizer",
8
  "unk_token": "<|endoftext|>"
9
  }