add train
Browse files- binhvq_dedup.py +4 -3
- train.jsonl.zst +3 -0
binhvq_dedup.py
CHANGED
@@ -23,7 +23,8 @@ Preprocessed:
|
|
23 |
|
24 |
_REPO_URL = "https://huggingface.co/datasets/imthanhlv/binhvq_dedup/tree/main/"
|
25 |
_URLS = {
|
26 |
-
"
|
|
|
27 |
}
|
28 |
|
29 |
|
@@ -62,7 +63,7 @@ class Binhvq(datasets.GeneratorBasedBuilder):
|
|
62 |
return [
|
63 |
datasets.SplitGenerator(
|
64 |
name=datasets.Split.TRAIN,
|
65 |
-
gen_kwargs={"filepath": downloaded_files["
|
66 |
),
|
67 |
datasets.SplitGenerator(
|
68 |
name=datasets.Split.VALIDATION,
|
@@ -74,6 +75,6 @@ class Binhvq(datasets.GeneratorBasedBuilder):
|
|
74 |
logging.warning(f"Generating examples from {filepath}")
|
75 |
_id = 0
|
76 |
reader = lm_dataformat.Reader(filepath)
|
77 |
-
for doc in
|
78 |
yield _id, {"text": doc}
|
79 |
_id += 1
|
|
|
23 |
|
24 |
_REPO_URL = "https://huggingface.co/datasets/imthanhlv/binhvq_dedup/tree/main/"
|
25 |
_URLS = {
|
26 |
+
"train" "train.jsonl.zst",
|
27 |
+
"val": "val.jsonl.zst"
|
28 |
}
|
29 |
|
30 |
|
|
|
63 |
return [
|
64 |
datasets.SplitGenerator(
|
65 |
name=datasets.Split.TRAIN,
|
66 |
+
gen_kwargs={"filepath": downloaded_files["train"]},
|
67 |
),
|
68 |
datasets.SplitGenerator(
|
69 |
name=datasets.Split.VALIDATION,
|
|
|
75 |
logging.warning(f"Generating examples from {filepath}")
|
76 |
_id = 0
|
77 |
reader = lm_dataformat.Reader(filepath)
|
78 |
+
for doc in reader.read_jsonl(filepath):
|
79 |
yield _id, {"text": doc}
|
80 |
_id += 1
|
train.jsonl.zst
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:add7eebdc28dae0edcbc9b3e51f06469870b573023fabce0703e6d388c4a218b
|
3 |
+
size 26841447213
|