Crystalcareai commited on
Commit
92aa7d1
·
1 Parent(s): fc142ce

Upload 5 files

Browse files
Files changed (3) hide show
  1. config.json +2 -5
  2. pytorch_model.bin +2 -2
  3. tokenizer_config.json +1 -1
config.json CHANGED
@@ -1,10 +1,8 @@
1
  {
2
- "_name_or_path": "/future/u/okhattab/root/unit/experiments/2021.10/none/kld.v3.nway32.l2.ib/checkpoints/colbert-150000/",
3
  "architectures": [
4
- "HF_ColBERT"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
- "classifier_dropout": null,
8
  "gradient_checkpointing": false,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
@@ -18,8 +16,7 @@
18
  "num_hidden_layers": 12,
19
  "pad_token_id": 0,
20
  "position_embedding_type": "absolute",
21
- "torch_dtype": "float32",
22
- "transformers_version": "4.10.0",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 30522
 
1
  {
 
2
  "architectures": [
3
+ "BertModel"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
6
  "gradient_checkpointing": false,
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
 
16
  "num_hidden_layers": 12,
17
  "pad_token_id": 0,
18
  "position_embedding_type": "absolute",
19
+ "transformers_version": "4.6.1",
 
20
  "type_vocab_size": 2,
21
  "use_cache": true,
22
  "vocab_size": 30522
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26e4c2f9f95a3da4442252bb40d99e4fbfd098e733edac1d785c9937b8a278da
3
- size 438405935
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:741908cf34398b6de142cb5bca8b1d1dc8dcafca0b4b24649a6ac456abdab0c8
3
+ size 438012727
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "/future/u/okhattab/root/unit/experiments/2021.10/none/kld.v3.nway32.l2.ib/checkpoints/colbert-150000/", "tokenizer_class": "BertTokenizer"}
 
1
+ {"do_lower_case": true, "do_basic_tokenize": true, "never_split": null, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased"}