mattdangerw commited on
Commit
da04892
·
verified ·
1 Parent(s): ae5a6ee

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. config.json +2 -4
  2. metadata.json +7 -3
  3. tokenizer.json +18 -9
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "module": "keras_nlp.src.models.roberta.roberta_backbone",
3
  "class_name": "RobertaBackbone",
4
  "config": {
5
  "name": "roberta_backbone",
@@ -12,7 +12,5 @@
12
  "dropout": 0.1,
13
  "max_sequence_length": 512
14
  },
15
- "registered_name": "keras_nlp>RobertaBackbone",
16
- "assets": [],
17
- "weights": "model.weights.h5"
18
  }
 
1
  {
2
+ "module": "keras_hub.src.models.roberta.roberta_backbone",
3
  "class_name": "RobertaBackbone",
4
  "config": {
5
  "name": "roberta_backbone",
 
12
  "dropout": 0.1,
13
  "max_sequence_length": 512
14
  },
15
+ "registered_name": "keras_hub>RobertaBackbone"
 
 
16
  }
metadata.json CHANGED
@@ -1,6 +1,10 @@
1
  {
2
- "keras_version": "3.0.1",
3
- "keras_nlp_version": "0.7.0",
4
  "parameter_count": 124052736,
5
- "date_saved": "2023-12-27@02:26:30"
 
 
 
 
6
  }
 
1
  {
2
+ "keras_version": "3.7.0",
3
+ "keras_hub_version": "0.19.0",
4
  "parameter_count": 124052736,
5
+ "date_saved": "2024-12-21@22:14:23",
6
+ "tasks": [
7
+ "MaskedLM",
8
+ "TextClassifier"
9
+ ]
10
  }
tokenizer.json CHANGED
@@ -1,17 +1,26 @@
1
  {
2
- "module": "keras_nlp.src.models.roberta.roberta_tokenizer",
3
  "class_name": "RobertaTokenizer",
4
  "config": {
5
  "name": "roberta_tokenizer",
6
  "trainable": true,
7
- "dtype": "int32",
 
 
 
 
 
 
 
 
8
  "sequence_length": null,
9
- "add_prefix_space": false
 
 
 
 
 
 
10
  },
11
- "registered_name": "keras_nlp>RobertaTokenizer",
12
- "assets": [
13
- "assets/tokenizer/merges.txt",
14
- "assets/tokenizer/vocabulary.json"
15
- ],
16
- "weights": null
17
  }
 
1
  {
2
+ "module": "keras_hub.src.models.roberta.roberta_tokenizer",
3
  "class_name": "RobertaTokenizer",
4
  "config": {
5
  "name": "roberta_tokenizer",
6
  "trainable": true,
7
+ "dtype": {
8
+ "module": "keras",
9
+ "class_name": "DTypePolicy",
10
+ "config": {
11
+ "name": "int32"
12
+ },
13
+ "registered_name": null
14
+ },
15
+ "config_file": "tokenizer.json",
16
  "sequence_length": null,
17
+ "add_prefix_space": false,
18
+ "unsplittable_tokens": [
19
+ "<s>",
20
+ "<pad>",
21
+ "</s>",
22
+ "<mask>"
23
+ ]
24
  },
25
+ "registered_name": "keras_hub>RobertaTokenizer"
 
 
 
 
 
26
  }