Datasets:
Tasks:
Other
Modalities:
Text
Languages:
Slovak
Size:
10K - 100K
Tags:
structure-prediction
License:
fix infos.json
Browse files
conll2003-sk-ner.py → conll2003-SK-NER.py
RENAMED
@@ -14,7 +14,7 @@
|
|
14 |
|
15 |
# Lint as: python3
|
16 |
"""Introduction to the CoNLL-2003-SK-NER Shared Task: Slovak Named Entity Recognition"""
|
17 |
-
|
18 |
import datasets
|
19 |
import json
|
20 |
|
@@ -28,8 +28,7 @@ _DESCRIPTION = """\
|
|
28 |
This is translated version of the original CONLL2003 dataset (translated from English to Slovak via Google translate) Annotation was done mostly automatically with word matching scripts. Records where some tags were not matched, were annotated manually (10%) Unlike the original Conll2003 dataset, this one contains only NER tags
|
29 |
"""
|
30 |
|
31 |
-
_URL="
|
32 |
-
#_URL = "https://github.com/ju-bezdek/conll2003-sk-ner/raw/master/data/"
|
33 |
_TRAINING_FILE = "train.json"
|
34 |
_DEV_FILE = "valid.json"
|
35 |
_TEST_FILE = "test.json"
|
@@ -80,14 +79,14 @@ class Conll2003(datasets.GeneratorBasedBuilder):
|
|
80 |
"""Returns SplitGenerators."""
|
81 |
urls_to_download = {
|
82 |
"train": f"{_URL}{_TRAINING_FILE}",
|
83 |
-
"
|
84 |
"test": f"{_URL}{_TEST_FILE}",
|
85 |
}
|
86 |
downloaded_files = dl_manager.download_and_extract(urls_to_download)
|
87 |
|
88 |
return [
|
89 |
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_files["train"]}),
|
90 |
-
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": downloaded_files["
|
91 |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files["test"]}),
|
92 |
]
|
93 |
|
|
|
14 |
|
15 |
# Lint as: python3
|
16 |
"""Introduction to the CoNLL-2003-SK-NER Shared Task: Slovak Named Entity Recognition"""
|
17 |
+
print("been here")
|
18 |
import datasets
|
19 |
import json
|
20 |
|
|
|
28 |
This is translated version of the original CONLL2003 dataset (translated from English to Slovak via Google translate) Annotation was done mostly automatically with word matching scripts. Records where some tags were not matched, were annotated manually (10%) Unlike the original Conll2003 dataset, this one contains only NER tags
|
29 |
"""
|
30 |
|
31 |
+
_URL="./data/"
|
|
|
32 |
_TRAINING_FILE = "train.json"
|
33 |
_DEV_FILE = "valid.json"
|
34 |
_TEST_FILE = "test.json"
|
|
|
79 |
"""Returns SplitGenerators."""
|
80 |
urls_to_download = {
|
81 |
"train": f"{_URL}{_TRAINING_FILE}",
|
82 |
+
"val": f"{_URL}{_DEV_FILE}",
|
83 |
"test": f"{_URL}{_TEST_FILE}",
|
84 |
}
|
85 |
downloaded_files = dl_manager.download_and_extract(urls_to_download)
|
86 |
|
87 |
return [
|
88 |
datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"filepath": downloaded_files["train"]}),
|
89 |
+
datasets.SplitGenerator(name=datasets.Split.VALIDATION, gen_kwargs={"filepath": downloaded_files["val"]}),
|
90 |
datasets.SplitGenerator(name=datasets.Split.TEST, gen_kwargs={"filepath": downloaded_files["test"]}),
|
91 |
]
|
92 |
|
dataset_infos.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"conll2003-SK-NER": {"description": "This is translated version of the original CONLL2003 dataset (translated from English to Slovak via Google translate) Annotation was done mostly automatically with word matching scripts. Records where some tags were not matched, were annotated manually (10%) Unlike the original Conll2003 dataset, this one contains only NER tags\n", "citation": "", "homepage": "", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-MISC", "I-MISC"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "conll2003", "config_name": "conll2003-SK-NER", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3333648, "num_examples": 14041, "dataset_name": "conll2003"}, "validation": {"name": "validation", "num_bytes": 927447, "num_examples": 3250, "dataset_name": "conll2003"}, "test": {"name": "test", "num_bytes": 853542, "num_examples": 3453, "dataset_name": "conll2003"}}, "download_checksums": {"
|
|
|
1 |
+
{"conll2003-SK-NER": {"description": "This is translated version of the original CONLL2003 dataset (translated from English to Slovak via Google translate) Annotation was done mostly automatically with word matching scripts. Records where some tags were not matched, were annotated manually (10%) Unlike the original Conll2003 dataset, this one contains only NER tags\n", "citation": "", "homepage": "", "license": "", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "ner_tags": {"feature": {"num_classes": 9, "names": ["O", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC", "B-MISC", "I-MISC"], "names_file": null, "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "conll2003", "config_name": "conll2003-SK-NER", "version": {"version_str": "1.0.0", "description": null, "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 3333648, "num_examples": 14041, "dataset_name": "conll2003"}, "validation": {"name": "validation", "num_bytes": 927447, "num_examples": 3250, "dataset_name": "conll2003"}, "test": {"name": "test", "num_bytes": 853542, "num_examples": 3453, "dataset_name": "conll2003"}}, "download_checksums": {"./data/train.json": {"num_bytes": 2680237, "checksum": "f4fe71debdfc0db1f3485bbcdb97b84086711956e235b5cd2c6d9a82fdfc5db6"}, "./data/valid.json": {"num_bytes": 702155, "checksum": "ea36d887933c6b0420399f940a878fbcc49b190ff4e1cade785c700c59d2ec5c"}, "./data/test.json": {"num_bytes": 650684, "checksum": "b0280959ea836976f2f0c8db8fdf5cb1b680f5e872cc4b9716339de8dc2a02ff"}}, "download_size": 4033076, "post_processing_size": null, "dataset_size": 5114637, "size_in_bytes": 9147713}}
|