Commit
•
c23fdff
1
Parent(s):
ffaa9c4
Convert dataset to Parquet (#5)
Browse files- Convert dataset to Parquet (064fab99e6683afedddfd4c01239a7ddae8c7266)
- Add ecthr_b data files (352f7328c0b977cc15c0abefadfe17686e934067)
- Add eurlex data files (d3e022f0dc4d2ccc3e39f221b0fd20606860655e)
- Add scotus data files (ffa3e5fdda2e9fc1b5a254ff46e0578440207e02)
- Add ledgar data files (6af614c5ab6d013a02f134e9e9f95373885327e3)
- Add unfair_tos data files (b7b8b5847681813d675d3945a80d2f1b3554be7f)
- Add case_hold data files (fe9d5f8f254c32c0d9c343e86534523deb2fcb01)
- Delete loading script (9c205f32915297e23b9922ab44df1a931823d7fd)
- Delete legacy dataset_infos.json (7ada00fe09504e4c6e7131da2dc8b2684921471b)
- README.md +149 -92
- case_hold/test-00000-of-00001.parquet +3 -0
- case_hold/train-00000-of-00001.parquet +3 -0
- case_hold/validation-00000-of-00001.parquet +3 -0
- dataset_infos.json +0 -1
- ecthr_a/test-00000-of-00001.parquet +3 -0
- ecthr_a/train-00000-of-00001.parquet +3 -0
- ecthr_a/validation-00000-of-00001.parquet +3 -0
- ecthr_b/test-00000-of-00001.parquet +3 -0
- ecthr_b/train-00000-of-00001.parquet +3 -0
- ecthr_b/validation-00000-of-00001.parquet +3 -0
- eurlex/test-00000-of-00001.parquet +3 -0
- eurlex/train-00000-of-00001.parquet +3 -0
- eurlex/validation-00000-of-00001.parquet +3 -0
- ledgar/test-00000-of-00001.parquet +3 -0
- ledgar/train-00000-of-00001.parquet +3 -0
- ledgar/validation-00000-of-00001.parquet +3 -0
- lex_glue.py +0 -659
- scotus/test-00000-of-00001.parquet +3 -0
- scotus/train-00000-of-00001.parquet +3 -0
- scotus/validation-00000-of-00001.parquet +3 -0
- unfair_tos/test-00000-of-00001.parquet +3 -0
- unfair_tos/train-00000-of-00001.parquet +3 -0
- unfair_tos/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -22,7 +22,42 @@ task_ids:
|
|
22 |
- multiple-choice-qa
|
23 |
- topic-classification
|
24 |
pretty_name: LexGLUE
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
25 |
dataset_info:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
- config_name: ecthr_a
|
27 |
features:
|
28 |
- name: text
|
@@ -43,16 +78,16 @@ dataset_info:
|
|
43 |
'9': P1-1
|
44 |
splits:
|
45 |
- name: train
|
46 |
-
num_bytes:
|
47 |
num_examples: 9000
|
48 |
- name: test
|
49 |
-
num_bytes:
|
50 |
num_examples: 1000
|
51 |
- name: validation
|
52 |
-
num_bytes:
|
53 |
num_examples: 1000
|
54 |
-
download_size:
|
55 |
-
dataset_size:
|
56 |
- config_name: ecthr_b
|
57 |
features:
|
58 |
- name: text
|
@@ -73,16 +108,16 @@ dataset_info:
|
|
73 |
'9': P1-1
|
74 |
splits:
|
75 |
- name: train
|
76 |
-
num_bytes:
|
77 |
num_examples: 9000
|
78 |
- name: test
|
79 |
-
num_bytes:
|
80 |
num_examples: 1000
|
81 |
- name: validation
|
82 |
-
num_bytes:
|
83 |
num_examples: 1000
|
84 |
-
download_size:
|
85 |
-
dataset_size:
|
86 |
- config_name: eurlex
|
87 |
features:
|
88 |
- name: text
|
@@ -193,49 +228,16 @@ dataset_info:
|
|
193 |
'99': '100285'
|
194 |
splits:
|
195 |
- name: train
|
196 |
-
num_bytes:
|
197 |
num_examples: 55000
|
198 |
- name: test
|
199 |
-
num_bytes:
|
200 |
num_examples: 5000
|
201 |
- name: validation
|
202 |
-
num_bytes:
|
203 |
num_examples: 5000
|
204 |
-
download_size:
|
205 |
-
dataset_size:
|
206 |
-
- config_name: scotus
|
207 |
-
features:
|
208 |
-
- name: text
|
209 |
-
dtype: string
|
210 |
-
- name: label
|
211 |
-
dtype:
|
212 |
-
class_label:
|
213 |
-
names:
|
214 |
-
'0': '1'
|
215 |
-
'1': '2'
|
216 |
-
'2': '3'
|
217 |
-
'3': '4'
|
218 |
-
'4': '5'
|
219 |
-
'5': '6'
|
220 |
-
'6': '7'
|
221 |
-
'7': '8'
|
222 |
-
'8': '9'
|
223 |
-
'9': '10'
|
224 |
-
'10': '11'
|
225 |
-
'11': '12'
|
226 |
-
'12': '13'
|
227 |
-
splits:
|
228 |
-
- name: train
|
229 |
-
num_bytes: 178959320
|
230 |
-
num_examples: 5000
|
231 |
-
- name: test
|
232 |
-
num_bytes: 76213283
|
233 |
-
num_examples: 1400
|
234 |
-
- name: validation
|
235 |
-
num_bytes: 75600247
|
236 |
-
num_examples: 1400
|
237 |
-
download_size: 104763335
|
238 |
-
dataset_size: 330772850
|
239 |
- config_name: ledgar
|
240 |
features:
|
241 |
- name: text
|
@@ -346,16 +348,49 @@ dataset_info:
|
|
346 |
'99': Withholdings
|
347 |
splits:
|
348 |
- name: train
|
349 |
-
num_bytes:
|
350 |
num_examples: 60000
|
351 |
- name: test
|
352 |
-
num_bytes:
|
353 |
num_examples: 10000
|
354 |
- name: validation
|
355 |
-
num_bytes:
|
356 |
num_examples: 10000
|
357 |
-
download_size:
|
358 |
-
dataset_size:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
359 |
- config_name: unfair_tos
|
360 |
features:
|
361 |
- name: text
|
@@ -374,51 +409,73 @@ dataset_info:
|
|
374 |
'7': Arbitration
|
375 |
splits:
|
376 |
- name: train
|
377 |
-
num_bytes:
|
378 |
num_examples: 5532
|
379 |
- name: test
|
380 |
-
num_bytes:
|
381 |
num_examples: 1607
|
382 |
- name: validation
|
383 |
-
num_bytes:
|
384 |
num_examples: 2275
|
385 |
-
download_size:
|
386 |
-
dataset_size:
|
|
|
387 |
- config_name: case_hold
|
388 |
-
|
389 |
-
-
|
390 |
-
|
391 |
-
-
|
392 |
-
|
393 |
-
-
|
394 |
-
|
395 |
-
|
396 |
-
|
397 |
-
|
398 |
-
|
399 |
-
|
400 |
-
|
401 |
-
|
402 |
-
|
403 |
-
|
404 |
-
|
405 |
-
|
406 |
-
|
407 |
-
|
408 |
-
|
409 |
-
-
|
410 |
-
|
411 |
-
|
412 |
-
|
413 |
-
|
414 |
-
|
415 |
-
-
|
416 |
-
|
417 |
-
-
|
418 |
-
|
419 |
-
- ledgar
|
420 |
-
|
421 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
422 |
---
|
423 |
|
424 |
# Dataset Card for "LexGLUE"
|
|
|
22 |
- multiple-choice-qa
|
23 |
- topic-classification
|
24 |
pretty_name: LexGLUE
|
25 |
+
config_names:
|
26 |
+
- case_hold
|
27 |
+
- ecthr_a
|
28 |
+
- ecthr_b
|
29 |
+
- eurlex
|
30 |
+
- ledgar
|
31 |
+
- scotus
|
32 |
+
- unfair_tos
|
33 |
dataset_info:
|
34 |
+
- config_name: case_hold
|
35 |
+
features:
|
36 |
+
- name: context
|
37 |
+
dtype: string
|
38 |
+
- name: endings
|
39 |
+
sequence: string
|
40 |
+
- name: label
|
41 |
+
dtype:
|
42 |
+
class_label:
|
43 |
+
names:
|
44 |
+
'0': '0'
|
45 |
+
'1': '1'
|
46 |
+
'2': '2'
|
47 |
+
'3': '3'
|
48 |
+
'4': '4'
|
49 |
+
splits:
|
50 |
+
- name: train
|
51 |
+
num_bytes: 74781706
|
52 |
+
num_examples: 45000
|
53 |
+
- name: test
|
54 |
+
num_bytes: 5989952
|
55 |
+
num_examples: 3600
|
56 |
+
- name: validation
|
57 |
+
num_bytes: 6474603
|
58 |
+
num_examples: 3900
|
59 |
+
download_size: 47303537
|
60 |
+
dataset_size: 87246261
|
61 |
- config_name: ecthr_a
|
62 |
features:
|
63 |
- name: text
|
|
|
78 |
'9': P1-1
|
79 |
splits:
|
80 |
- name: train
|
81 |
+
num_bytes: 89637449
|
82 |
num_examples: 9000
|
83 |
- name: test
|
84 |
+
num_bytes: 11884168
|
85 |
num_examples: 1000
|
86 |
- name: validation
|
87 |
+
num_bytes: 10985168
|
88 |
num_examples: 1000
|
89 |
+
download_size: 53352586
|
90 |
+
dataset_size: 112506785
|
91 |
- config_name: ecthr_b
|
92 |
features:
|
93 |
- name: text
|
|
|
108 |
'9': P1-1
|
109 |
splits:
|
110 |
- name: train
|
111 |
+
num_bytes: 89657649
|
112 |
num_examples: 9000
|
113 |
- name: test
|
114 |
+
num_bytes: 11886928
|
115 |
num_examples: 1000
|
116 |
- name: validation
|
117 |
+
num_bytes: 10987816
|
118 |
num_examples: 1000
|
119 |
+
download_size: 53352494
|
120 |
+
dataset_size: 112532393
|
121 |
- config_name: eurlex
|
122 |
features:
|
123 |
- name: text
|
|
|
228 |
'99': '100285'
|
229 |
splits:
|
230 |
- name: train
|
231 |
+
num_bytes: 390770241
|
232 |
num_examples: 55000
|
233 |
- name: test
|
234 |
+
num_bytes: 59739094
|
235 |
num_examples: 5000
|
236 |
- name: validation
|
237 |
+
num_bytes: 41544476
|
238 |
num_examples: 5000
|
239 |
+
download_size: 208028049
|
240 |
+
dataset_size: 492053811
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
241 |
- config_name: ledgar
|
242 |
features:
|
243 |
- name: text
|
|
|
348 |
'99': Withholdings
|
349 |
splits:
|
350 |
- name: train
|
351 |
+
num_bytes: 43358291
|
352 |
num_examples: 60000
|
353 |
- name: test
|
354 |
+
num_bytes: 6845581
|
355 |
num_examples: 10000
|
356 |
- name: validation
|
357 |
+
num_bytes: 7143588
|
358 |
num_examples: 10000
|
359 |
+
download_size: 27650585
|
360 |
+
dataset_size: 57347460
|
361 |
+
- config_name: scotus
|
362 |
+
features:
|
363 |
+
- name: text
|
364 |
+
dtype: string
|
365 |
+
- name: label
|
366 |
+
dtype:
|
367 |
+
class_label:
|
368 |
+
names:
|
369 |
+
'0': '1'
|
370 |
+
'1': '2'
|
371 |
+
'2': '3'
|
372 |
+
'3': '4'
|
373 |
+
'4': '5'
|
374 |
+
'5': '6'
|
375 |
+
'6': '7'
|
376 |
+
'7': '8'
|
377 |
+
'8': '9'
|
378 |
+
'9': '10'
|
379 |
+
'10': '11'
|
380 |
+
'11': '12'
|
381 |
+
'12': '13'
|
382 |
+
splits:
|
383 |
+
- name: train
|
384 |
+
num_bytes: 178959316
|
385 |
+
num_examples: 5000
|
386 |
+
- name: test
|
387 |
+
num_bytes: 76213279
|
388 |
+
num_examples: 1400
|
389 |
+
- name: validation
|
390 |
+
num_bytes: 75600243
|
391 |
+
num_examples: 1400
|
392 |
+
download_size: 173411399
|
393 |
+
dataset_size: 330772838
|
394 |
- config_name: unfair_tos
|
395 |
features:
|
396 |
- name: text
|
|
|
409 |
'7': Arbitration
|
410 |
splits:
|
411 |
- name: train
|
412 |
+
num_bytes: 1041782
|
413 |
num_examples: 5532
|
414 |
- name: test
|
415 |
+
num_bytes: 303099
|
416 |
num_examples: 1607
|
417 |
- name: validation
|
418 |
+
num_bytes: 452111
|
419 |
num_examples: 2275
|
420 |
+
download_size: 865604
|
421 |
+
dataset_size: 1796992
|
422 |
+
configs:
|
423 |
- config_name: case_hold
|
424 |
+
data_files:
|
425 |
+
- split: train
|
426 |
+
path: case_hold/train-*
|
427 |
+
- split: test
|
428 |
+
path: case_hold/test-*
|
429 |
+
- split: validation
|
430 |
+
path: case_hold/validation-*
|
431 |
+
- config_name: ecthr_a
|
432 |
+
data_files:
|
433 |
+
- split: train
|
434 |
+
path: ecthr_a/train-*
|
435 |
+
- split: test
|
436 |
+
path: ecthr_a/test-*
|
437 |
+
- split: validation
|
438 |
+
path: ecthr_a/validation-*
|
439 |
+
- config_name: ecthr_b
|
440 |
+
data_files:
|
441 |
+
- split: train
|
442 |
+
path: ecthr_b/train-*
|
443 |
+
- split: test
|
444 |
+
path: ecthr_b/test-*
|
445 |
+
- split: validation
|
446 |
+
path: ecthr_b/validation-*
|
447 |
+
- config_name: eurlex
|
448 |
+
data_files:
|
449 |
+
- split: train
|
450 |
+
path: eurlex/train-*
|
451 |
+
- split: test
|
452 |
+
path: eurlex/test-*
|
453 |
+
- split: validation
|
454 |
+
path: eurlex/validation-*
|
455 |
+
- config_name: ledgar
|
456 |
+
data_files:
|
457 |
+
- split: train
|
458 |
+
path: ledgar/train-*
|
459 |
+
- split: test
|
460 |
+
path: ledgar/test-*
|
461 |
+
- split: validation
|
462 |
+
path: ledgar/validation-*
|
463 |
+
- config_name: scotus
|
464 |
+
data_files:
|
465 |
+
- split: train
|
466 |
+
path: scotus/train-*
|
467 |
+
- split: test
|
468 |
+
path: scotus/test-*
|
469 |
+
- split: validation
|
470 |
+
path: scotus/validation-*
|
471 |
+
- config_name: unfair_tos
|
472 |
+
data_files:
|
473 |
+
- split: train
|
474 |
+
path: unfair_tos/train-*
|
475 |
+
- split: test
|
476 |
+
path: unfair_tos/test-*
|
477 |
+
- split: validation
|
478 |
+
path: unfair_tos/validation-*
|
479 |
---
|
480 |
|
481 |
# Dataset Card for "LexGLUE"
|
case_hold/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0028b12e681c1ddd4370a53d6670334a4cf2c8b3179f1d5c2c9c10a95a1afe37
|
3 |
+
size 3256651
|
case_hold/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b0c4a7c5f5228d412a09411b6feac17cff791d072efb72d567dd5460ec7a5925
|
3 |
+
size 40533452
|
case_hold/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:94d17463cae99ffc9f86bb1faa033bab7b71e08f32cd83550ce96d4d71f2638d
|
3 |
+
size 3513434
|
dataset_infos.json
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
{"ecthr_a": {"description": "The European Court of Human Rights (ECtHR) hears allegations that a state has\nbreached human rights provisions of the European Convention of Human Rights (ECHR).\nFor each case, the dataset provides a list of factual paragraphs (facts) from the case description.\nEach case is mapped to articles of the ECHR that were violated (if any).", "citation": "@inproceedings{chalkidis-etal-2021-paragraph,\n title = \"Paragraph-level Rationale Extraction through Regularization: A case study on {E}uropean Court of Human Rights Cases\",\n author = \"Chalkidis, Ilias and\n Fergadiotis, Manos and\n Tsarapatsanis, Dimitrios and\n Aletras, Nikolaos and\n Androutsopoulos, Ion and\n Malakasiotis, Prodromos\",\n booktitle = \"Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n month = jun,\n year = \"2021\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://aclanthology.org/2021.naacl-main.22\",\n doi = \"10.18653/v1/2021.naacl-main.22\",\n pages = \"226--241\",\n}\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "https://archive.org/details/ECtHR-NAACL2021", "license": "", "features": {"text": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "labels": {"feature": {"num_classes": 10, "names": ["2", "3", "5", "6", "8", "9", "10", "11", "14", "P1-1"], "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "ecthr_a", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 89637461, "num_examples": 9000, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 11884180, "num_examples": 1000, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 10985180, "num_examples": 1000, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/ecthr.tar.gz": {"num_bytes": 32852475, "checksum": "461c1f6016af3a7ac0bd115c1f9ff65031258bfec39e570fec74a16d8946398e"}}, "download_size": 32852475, "post_processing_size": null, "dataset_size": 112506821, "size_in_bytes": 145359296}, "ecthr_b": {"description": "The European Court of Human Rights (ECtHR) hears allegations that a state has\nbreached human rights provisions of the European Convention of Human Rights (ECHR).\nFor each case, the dataset provides a list of factual paragraphs (facts) from the case description.\nEach case is mapped to articles of ECHR that were allegedly violated (considered by the court).", "citation": "@inproceedings{chalkidis-etal-2021-paragraph,\n title = \"Paragraph-level Rationale Extraction through Regularization: A case study on {E}uropean Court of Human Rights Cases\",\n author = \"Chalkidis, Ilias\n and Fergadiotis, Manos\n and Tsarapatsanis, Dimitrios\n and Aletras, Nikolaos\n and Androutsopoulos, Ion\n and Malakasiotis, Prodromos\",\n booktitle = \"Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies\",\n year = \"2021\",\n address = \"Online\",\n url = \"https://aclanthology.org/2021.naacl-main.22\",\n}\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "https://archive.org/details/ECtHR-NAACL2021", "license": "", "features": {"text": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "labels": {"feature": {"num_classes": 10, "names": ["2", "3", "5", "6", "8", "9", "10", "11", "14", "P1-1"], "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "ecthr_b", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 89657661, "num_examples": 9000, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 11886940, "num_examples": 1000, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 10987828, "num_examples": 1000, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/ecthr.tar.gz": {"num_bytes": 32852475, "checksum": "461c1f6016af3a7ac0bd115c1f9ff65031258bfec39e570fec74a16d8946398e"}}, "download_size": 32852475, "post_processing_size": null, "dataset_size": 112532429, "size_in_bytes": 145384904}, "eurlex": {"description": "European Union (EU) legislation is published in EUR-Lex portal.\nAll EU laws are annotated by EU's Publications Office with multiple concepts from the EuroVoc thesaurus,\na multilingual thesaurus maintained by the Publications Office.\nThe current version of EuroVoc contains more than 7k concepts referring to various activities\nof the EU and its Member States (e.g., economics, health-care, trade).\nGiven a document, the task is to predict its EuroVoc labels (concepts).", "citation": "@inproceedings{chalkidis-etal-2021-multieurlex,\n author = {Chalkidis, Ilias and\n Fergadiotis, Manos and\n Androutsopoulos, Ion},\n title = {MultiEURLEX -- A multi-lingual and multi-label legal document\n classification dataset for zero-shot cross-lingual transfer},\n booktitle = {Proceedings of the 2021 Conference on Empirical Methods\n in Natural Language Processing},\n year = {2021},\n location = {Punta Cana, Dominican Republic},\n}\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "https://zenodo.org/record/5363165#.YVJOAi8RqaA", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "labels": {"feature": {"num_classes": 100, "names": ["100163", "100168", "100169", "100170", "100171", "100172", "100173", "100174", "100175", "100176", "100177", "100179", "100180", "100183", "100184", "100185", "100186", "100187", "100189", "100190", "100191", "100192", "100193", "100194", "100195", "100196", "100197", "100198", "100199", "100200", "100201", "100202", "100204", "100205", "100206", "100207", "100212", "100214", "100215", "100220", "100221", "100222", "100223", "100224", "100226", "100227", "100229", "100230", "100231", "100232", "100233", "100234", "100235", "100237", "100238", "100239", "100240", "100241", "100242", "100243", "100244", "100245", "100246", "100247", "100248", "100249", "100250", "100252", "100253", "100254", "100255", "100256", "100257", "100258", "100259", "100260", "100261", "100262", "100263", "100264", "100265", "100266", "100268", "100269", "100270", "100271", "100272", "100273", "100274", "100275", "100276", "100277", "100278", "100279", "100280", "100281", "100282", "100283", "100284", "100285"], "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "eurlex", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 390770289, "num_examples": 55000, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 59739102, "num_examples": 5000, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 41544484, "num_examples": 5000, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/eurlex.tar.gz": {"num_bytes": 125413277, "checksum": "82376ff55c3812632d8a21ad0d7e515e2e7ec6431ca7673a454cdd41a3a7bf46"}}, "download_size": 125413277, "post_processing_size": null, "dataset_size": 492053875, "size_in_bytes": 617467152}, "scotus": {"description": "The US Supreme Court (SCOTUS) is the highest federal court in the United States of America\nand generally hears only the most controversial or otherwise complex cases which have not\nbeen sufficiently well solved by lower courts. This is a single-label multi-class classification\ntask, where given a document (court opinion), the task is to predict the relevant issue areas.\nThe 14 issue areas cluster 278 issues whose focus is on the subject matter of the controversy (dispute).", "citation": "@misc{spaeth2020,\n author = {Harold J. Spaeth and Lee Epstein and Andrew D. Martin, Jeffrey A. Segal\n and Theodore J. Ruger and Sara C. Benesh},\n year = {2020},\n title ={{Supreme Court Database, Version 2020 Release 01}},\n url= {http://Supremecourtdatabase.org},\n howpublished={Washington University Law}\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "http://scdb.wustl.edu/data.php", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 13, "names": ["1", "2", "3", "4", "5", "6", "7", "8", "9", "10", "11", "12", "13"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "scotus", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 178959320, "num_examples": 5000, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 76213283, "num_examples": 1400, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 75600247, "num_examples": 1400, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/scotus.tar.gz": {"num_bytes": 104763335, "checksum": "d53cc99aaf60b24ca7e4cf634f08a2572b5b3532f83aecdfc2c4257050dc9d0a"}}, "download_size": 104763335, "post_processing_size": null, "dataset_size": 330772850, "size_in_bytes": 435536185}, "ledgar": {"description": "LEDGAR dataset aims contract provision (paragraph) classification.\nThe contract provisions come from contracts obtained from the US Securities and Exchange Commission (SEC)\nfilings, which are publicly available from EDGAR. Each label represents the single main topic\n(theme) of the corresponding contract provision.", "citation": "@inproceedings{tuggener-etal-2020-ledgar,\n title = \"{LEDGAR}: A Large-Scale Multi-label Corpus for Text Classification of Legal Provisions in Contracts\",\n author = {Tuggener, Don and\n von D{\"a}niken, Pius and\n Peetz, Thomas and\n Cieliebak, Mark},\n booktitle = \"Proceedings of the 12th Language Resources and Evaluation Conference\",\n year = \"2020\",\n address = \"Marseille, France\",\n url = \"https://aclanthology.org/2020.lrec-1.155\",\n}\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "https://metatext.io/datasets/ledgar", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 100, "names": ["Adjustments", "Agreements", "Amendments", "Anti-Corruption Laws", "Applicable Laws", "Approvals", "Arbitration", "Assignments", "Assigns", "Authority", "Authorizations", "Base Salary", "Benefits", "Binding Effects", "Books", "Brokers", "Capitalization", "Change In Control", "Closings", "Compliance With Laws", "Confidentiality", "Consent To Jurisdiction", "Consents", "Construction", "Cooperation", "Costs", "Counterparts", "Death", "Defined Terms", "Definitions", "Disability", "Disclosures", "Duties", "Effective Dates", "Effectiveness", "Employment", "Enforceability", "Enforcements", "Entire Agreements", "Erisa", "Existence", "Expenses", "Fees", "Financial Statements", "Forfeitures", "Further Assurances", "General", "Governing Laws", "Headings", "Indemnifications", "Indemnity", "Insurances", "Integration", "Intellectual Property", "Interests", "Interpretations", "Jurisdictions", "Liens", "Litigations", "Miscellaneous", "Modifications", "No Conflicts", "No Defaults", "No Waivers", "Non-Disparagement", "Notices", "Organizations", "Participations", "Payments", "Positions", "Powers", "Publicity", "Qualifications", "Records", "Releases", "Remedies", "Representations", "Sales", "Sanctions", "Severability", "Solvency", "Specific Performance", "Submission To Jurisdiction", "Subsidiaries", "Successors", "Survival", "Tax Withholdings", "Taxes", "Terminations", "Terms", "Titles", "Transactions With Affiliates", "Use Of Proceeds", "Vacations", "Venues", "Vesting", "Waiver Of Jury Trials", "Waivers", "Warranties", "Withholdings"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "ledgar", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 43358315, "num_examples": 60000, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 6845585, "num_examples": 10000, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 7143592, "num_examples": 10000, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/ledgar.tar.gz": {"num_bytes": 16255623, "checksum": "f7507bcce46ce03e3e91b8aaa1b84ddf6e8f1d628c0d7fa351f97ce45366d5d8"}}, "download_size": 16255623, "post_processing_size": null, "dataset_size": 57347492, "size_in_bytes": 73603115}, "unfair_tos": {"description": "The UNFAIR-ToS dataset contains 50 Terms of Service (ToS) from on-line platforms (e.g., YouTube,\nEbay, Facebook, etc.). The dataset has been annotated on the sentence-level with 8 types of\nunfair contractual terms (sentences), meaning terms that potentially violate user rights\naccording to the European consumer law.", "citation": "@article{lippi-etal-2019-claudette,\n title = \"{CLAUDETTE}: an automated detector of potentially unfair clauses in online terms of service\",\n author = {Lippi, Marco\n and Pa\u0142ka, Przemys\u0142aw\n and Contissa, Giuseppe\n and Lagioia, Francesca\n and Micklitz, Hans-Wolfgang\n and Sartor, Giovanni\n and Torroni, Paolo},\n journal = \"Artificial Intelligence and Law\",\n year = \"2019\",\n publisher = \"Springer\",\n url = \"https://doi.org/10.1007/s10506-019-09243-2\",\n pages = \"117--139\",\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "http://claudette.eui.eu", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "labels": {"feature": {"num_classes": 8, "names": ["Limitation of liability", "Unilateral termination", "Unilateral change", "Content removal", "Contract by using", "Choice of law", "Jurisdiction", "Arbitration"], "id": null, "_type": "ClassLabel"}, "length": -1, "id": null, "_type": "Sequence"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "unfair_tos", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 1041790, "num_examples": 5532, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 303107, "num_examples": 1607, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 452119, "num_examples": 2275, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/unfair_tos.tar.gz": {"num_bytes": 511342, "checksum": "934470d74b62139dfbfad4a13b75a32e4a4d26a680ab12eedfb7659cdf669d53"}}, "download_size": 511342, "post_processing_size": null, "dataset_size": 1797016, "size_in_bytes": 2308358}, "case_hold": {"description": "The CaseHOLD (Case Holdings on Legal Decisions) dataset contains approx. 53k multiple choice\nquestions about holdings of US court cases from the Harvard Law Library case law corpus.\nHoldings are short summaries of legal rulings accompany referenced decisions relevant for the present case.\nThe input consists of an excerpt (or prompt) from a court decision, containing a reference\nto a particular case, while the holding statement is masked out. The model must identify\nthe correct (masked) holding statement from a selection of five choices.", "citation": "@inproceedings{Zheng2021,\n author = {Lucia Zheng and\n Neel Guha and\n Brandon R. Anderson and\n Peter Henderson and\n Daniel E. Ho},\n title = {When Does Pretraining Help? Assessing Self-Supervised Learning for\n Law and the CaseHOLD Dataset},\n year = {2021},\n booktitle = {International Conference on Artificial Intelligence and Law},\n}\n@article{chalkidis-etal-2021-lexglue,\n title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},\n author={Chalkidis, Ilias and\n Jana, Abhik and\n Hartung, Dirk and\n Bommarito, Michael and\n Androutsopoulos, Ion and\n Katz, Daniel Martin and\n Aletras, Nikolaos},\n year={2021},\n eprint={2110.00976},\n archivePrefix={arXiv},\n primaryClass={cs.CL},\n note = {arXiv: 2110.00976},\n}", "homepage": "https://github.com/reglab/casehold", "license": "", "features": {"context": {"dtype": "string", "id": null, "_type": "Value"}, "endings": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "label": {"num_classes": 5, "names": ["0", "1", "2", "3", "4"], "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "lex_glue", "config_name": "case_hold", "version": {"version_str": "1.0.0", "description": "", "major": 1, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 74781766, "num_examples": 45000, "dataset_name": "lex_glue"}, "test": {"name": "test", "num_bytes": 5989964, "num_examples": 3600, "dataset_name": "lex_glue"}, "validation": {"name": "validation", "num_bytes": 6474615, "num_examples": 3900, "dataset_name": "lex_glue"}}, "download_checksums": {"https://zenodo.org/record/5532997/files/casehold.tar.gz": {"num_bytes": 30422703, "checksum": "728827dae0019880fe6be609e23f8c47fa2b49a2f0814a36687ace8db1c32d5e"}}, "download_size": 30422703, "post_processing_size": null, "dataset_size": 87246345, "size_in_bytes": 117669048}}
|
|
|
|
ecthr_a/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e306383d3ffd3baec21ba7419c03a9d2273f1c0946cf2d5512cf98c81d1a3f8b
|
3 |
+
size 5677114
|
ecthr_a/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2f263758eb80ec1c532e1c4bde6cbfb2a8ff41378ebec5999aca23929c39cdf4
|
3 |
+
size 42415324
|
ecthr_a/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d74de774cf566f7eea52563b54dec5721de6b7f5a34438c698aa48d7ff9de205
|
3 |
+
size 5260148
|
ecthr_b/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d41d3667819cce507b57f6a59da28fc3691a722b6cfb7f4bc74805f0e713df6e
|
3 |
+
size 5677091
|
ecthr_b/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49ccf9e24ec08a05d5c69f2405e15e39c8472c6f786d03857abf914055befff6
|
3 |
+
size 42415290
|
ecthr_b/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fce01b0c0c5886b48ddd1f16f69ee1c3b7b2a9a5e96d2f3a601fe68c832b67d4
|
3 |
+
size 5260113
|
eurlex/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0c2ce5e5bc26c907607613aa1d7e4209553dbc3b85ec6ddcff1c745e2c7fdb9d
|
3 |
+
size 24273776
|
eurlex/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c0370abfe067fadec918490a01ba43ffc5bfdc6155283baffa5c120cf7754d3a
|
3 |
+
size 166694290
|
eurlex/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7f15ce05817231e33b5e0d63f3c07972a0490e1caa2bbfed4d6316204c893c8
|
3 |
+
size 17059983
|
ledgar/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af9ed0ec09545d39449ffddca3996a1a7774e2bd0a9d6fc09d4e9dd067892a7a
|
3 |
+
size 3313508
|
ledgar/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2901d9e019f5652dadda839d488155c681e50b90bd33f00574949418827867d
|
3 |
+
size 20897973
|
ledgar/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55ef0deada1ed75bbf13c68fdd1c1a1ebfa5b09176d5e68b92499234f51d6c9f
|
3 |
+
size 3439104
|
lex_glue.py
DELETED
@@ -1,659 +0,0 @@
|
|
1 |
-
# coding=utf-8
|
2 |
-
# Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
|
3 |
-
#
|
4 |
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
5 |
-
# you may not use this file except in compliance with the License.
|
6 |
-
# You may obtain a copy of the License at
|
7 |
-
#
|
8 |
-
# http://www.apache.org/licenses/LICENSE-2.0
|
9 |
-
#
|
10 |
-
# Unless required by applicable law or agreed to in writing, software
|
11 |
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
12 |
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
13 |
-
# See the License for the specific language governing permissions and
|
14 |
-
# limitations under the License.
|
15 |
-
"""LexGLUE: A Benchmark Dataset for Legal Language Understanding in English."""
|
16 |
-
|
17 |
-
import csv
|
18 |
-
import json
|
19 |
-
import textwrap
|
20 |
-
|
21 |
-
import datasets
|
22 |
-
|
23 |
-
|
24 |
-
MAIN_CITATION = """\
|
25 |
-
@article{chalkidis-etal-2021-lexglue,
|
26 |
-
title={{LexGLUE}: A Benchmark Dataset for Legal Language Understanding in English},
|
27 |
-
author={Chalkidis, Ilias and
|
28 |
-
Jana, Abhik and
|
29 |
-
Hartung, Dirk and
|
30 |
-
Bommarito, Michael and
|
31 |
-
Androutsopoulos, Ion and
|
32 |
-
Katz, Daniel Martin and
|
33 |
-
Aletras, Nikolaos},
|
34 |
-
year={2021},
|
35 |
-
eprint={2110.00976},
|
36 |
-
archivePrefix={arXiv},
|
37 |
-
primaryClass={cs.CL},
|
38 |
-
note = {arXiv: 2110.00976},
|
39 |
-
}"""
|
40 |
-
|
41 |
-
_DESCRIPTION = """\
|
42 |
-
Legal General Language Understanding Evaluation (LexGLUE) benchmark is
|
43 |
-
a collection of datasets for evaluating model performance across a diverse set of legal NLU tasks
|
44 |
-
"""
|
45 |
-
|
46 |
-
ECTHR_ARTICLES = ["2", "3", "5", "6", "8", "9", "10", "11", "14", "P1-1"]
|
47 |
-
|
48 |
-
EUROVOC_CONCEPTS = [
|
49 |
-
"100163",
|
50 |
-
"100168",
|
51 |
-
"100169",
|
52 |
-
"100170",
|
53 |
-
"100171",
|
54 |
-
"100172",
|
55 |
-
"100173",
|
56 |
-
"100174",
|
57 |
-
"100175",
|
58 |
-
"100176",
|
59 |
-
"100177",
|
60 |
-
"100179",
|
61 |
-
"100180",
|
62 |
-
"100183",
|
63 |
-
"100184",
|
64 |
-
"100185",
|
65 |
-
"100186",
|
66 |
-
"100187",
|
67 |
-
"100189",
|
68 |
-
"100190",
|
69 |
-
"100191",
|
70 |
-
"100192",
|
71 |
-
"100193",
|
72 |
-
"100194",
|
73 |
-
"100195",
|
74 |
-
"100196",
|
75 |
-
"100197",
|
76 |
-
"100198",
|
77 |
-
"100199",
|
78 |
-
"100200",
|
79 |
-
"100201",
|
80 |
-
"100202",
|
81 |
-
"100204",
|
82 |
-
"100205",
|
83 |
-
"100206",
|
84 |
-
"100207",
|
85 |
-
"100212",
|
86 |
-
"100214",
|
87 |
-
"100215",
|
88 |
-
"100220",
|
89 |
-
"100221",
|
90 |
-
"100222",
|
91 |
-
"100223",
|
92 |
-
"100224",
|
93 |
-
"100226",
|
94 |
-
"100227",
|
95 |
-
"100229",
|
96 |
-
"100230",
|
97 |
-
"100231",
|
98 |
-
"100232",
|
99 |
-
"100233",
|
100 |
-
"100234",
|
101 |
-
"100235",
|
102 |
-
"100237",
|
103 |
-
"100238",
|
104 |
-
"100239",
|
105 |
-
"100240",
|
106 |
-
"100241",
|
107 |
-
"100242",
|
108 |
-
"100243",
|
109 |
-
"100244",
|
110 |
-
"100245",
|
111 |
-
"100246",
|
112 |
-
"100247",
|
113 |
-
"100248",
|
114 |
-
"100249",
|
115 |
-
"100250",
|
116 |
-
"100252",
|
117 |
-
"100253",
|
118 |
-
"100254",
|
119 |
-
"100255",
|
120 |
-
"100256",
|
121 |
-
"100257",
|
122 |
-
"100258",
|
123 |
-
"100259",
|
124 |
-
"100260",
|
125 |
-
"100261",
|
126 |
-
"100262",
|
127 |
-
"100263",
|
128 |
-
"100264",
|
129 |
-
"100265",
|
130 |
-
"100266",
|
131 |
-
"100268",
|
132 |
-
"100269",
|
133 |
-
"100270",
|
134 |
-
"100271",
|
135 |
-
"100272",
|
136 |
-
"100273",
|
137 |
-
"100274",
|
138 |
-
"100275",
|
139 |
-
"100276",
|
140 |
-
"100277",
|
141 |
-
"100278",
|
142 |
-
"100279",
|
143 |
-
"100280",
|
144 |
-
"100281",
|
145 |
-
"100282",
|
146 |
-
"100283",
|
147 |
-
"100284",
|
148 |
-
"100285",
|
149 |
-
]
|
150 |
-
|
151 |
-
LEDGAR_CATEGORIES = [
|
152 |
-
"Adjustments",
|
153 |
-
"Agreements",
|
154 |
-
"Amendments",
|
155 |
-
"Anti-Corruption Laws",
|
156 |
-
"Applicable Laws",
|
157 |
-
"Approvals",
|
158 |
-
"Arbitration",
|
159 |
-
"Assignments",
|
160 |
-
"Assigns",
|
161 |
-
"Authority",
|
162 |
-
"Authorizations",
|
163 |
-
"Base Salary",
|
164 |
-
"Benefits",
|
165 |
-
"Binding Effects",
|
166 |
-
"Books",
|
167 |
-
"Brokers",
|
168 |
-
"Capitalization",
|
169 |
-
"Change In Control",
|
170 |
-
"Closings",
|
171 |
-
"Compliance With Laws",
|
172 |
-
"Confidentiality",
|
173 |
-
"Consent To Jurisdiction",
|
174 |
-
"Consents",
|
175 |
-
"Construction",
|
176 |
-
"Cooperation",
|
177 |
-
"Costs",
|
178 |
-
"Counterparts",
|
179 |
-
"Death",
|
180 |
-
"Defined Terms",
|
181 |
-
"Definitions",
|
182 |
-
"Disability",
|
183 |
-
"Disclosures",
|
184 |
-
"Duties",
|
185 |
-
"Effective Dates",
|
186 |
-
"Effectiveness",
|
187 |
-
"Employment",
|
188 |
-
"Enforceability",
|
189 |
-
"Enforcements",
|
190 |
-
"Entire Agreements",
|
191 |
-
"Erisa",
|
192 |
-
"Existence",
|
193 |
-
"Expenses",
|
194 |
-
"Fees",
|
195 |
-
"Financial Statements",
|
196 |
-
"Forfeitures",
|
197 |
-
"Further Assurances",
|
198 |
-
"General",
|
199 |
-
"Governing Laws",
|
200 |
-
"Headings",
|
201 |
-
"Indemnifications",
|
202 |
-
"Indemnity",
|
203 |
-
"Insurances",
|
204 |
-
"Integration",
|
205 |
-
"Intellectual Property",
|
206 |
-
"Interests",
|
207 |
-
"Interpretations",
|
208 |
-
"Jurisdictions",
|
209 |
-
"Liens",
|
210 |
-
"Litigations",
|
211 |
-
"Miscellaneous",
|
212 |
-
"Modifications",
|
213 |
-
"No Conflicts",
|
214 |
-
"No Defaults",
|
215 |
-
"No Waivers",
|
216 |
-
"Non-Disparagement",
|
217 |
-
"Notices",
|
218 |
-
"Organizations",
|
219 |
-
"Participations",
|
220 |
-
"Payments",
|
221 |
-
"Positions",
|
222 |
-
"Powers",
|
223 |
-
"Publicity",
|
224 |
-
"Qualifications",
|
225 |
-
"Records",
|
226 |
-
"Releases",
|
227 |
-
"Remedies",
|
228 |
-
"Representations",
|
229 |
-
"Sales",
|
230 |
-
"Sanctions",
|
231 |
-
"Severability",
|
232 |
-
"Solvency",
|
233 |
-
"Specific Performance",
|
234 |
-
"Submission To Jurisdiction",
|
235 |
-
"Subsidiaries",
|
236 |
-
"Successors",
|
237 |
-
"Survival",
|
238 |
-
"Tax Withholdings",
|
239 |
-
"Taxes",
|
240 |
-
"Terminations",
|
241 |
-
"Terms",
|
242 |
-
"Titles",
|
243 |
-
"Transactions With Affiliates",
|
244 |
-
"Use Of Proceeds",
|
245 |
-
"Vacations",
|
246 |
-
"Venues",
|
247 |
-
"Vesting",
|
248 |
-
"Waiver Of Jury Trials",
|
249 |
-
"Waivers",
|
250 |
-
"Warranties",
|
251 |
-
"Withholdings",
|
252 |
-
]
|
253 |
-
|
254 |
-
SCDB_ISSUE_AREAS = ["1", "2", "3", "4", "5", "6", "7", "8", "9", "10", "11", "12", "13"]
|
255 |
-
|
256 |
-
UNFAIR_CATEGORIES = [
|
257 |
-
"Limitation of liability",
|
258 |
-
"Unilateral termination",
|
259 |
-
"Unilateral change",
|
260 |
-
"Content removal",
|
261 |
-
"Contract by using",
|
262 |
-
"Choice of law",
|
263 |
-
"Jurisdiction",
|
264 |
-
"Arbitration",
|
265 |
-
]
|
266 |
-
|
267 |
-
CASEHOLD_LABELS = ["0", "1", "2", "3", "4"]
|
268 |
-
|
269 |
-
|
270 |
-
class LexGlueConfig(datasets.BuilderConfig):
|
271 |
-
"""BuilderConfig for LexGLUE."""
|
272 |
-
|
273 |
-
def __init__(
|
274 |
-
self,
|
275 |
-
text_column,
|
276 |
-
label_column,
|
277 |
-
url,
|
278 |
-
data_url,
|
279 |
-
data_file,
|
280 |
-
citation,
|
281 |
-
label_classes=None,
|
282 |
-
multi_label=None,
|
283 |
-
dev_column="dev",
|
284 |
-
**kwargs,
|
285 |
-
):
|
286 |
-
"""BuilderConfig for LexGLUE.
|
287 |
-
|
288 |
-
Args:
|
289 |
-
text_column: ``string`, name of the column in the jsonl file corresponding
|
290 |
-
to the text
|
291 |
-
label_column: `string`, name of the column in the jsonl file corresponding
|
292 |
-
to the label
|
293 |
-
url: `string`, url for the original project
|
294 |
-
data_url: `string`, url to download the zip file from
|
295 |
-
data_file: `string`, filename for data set
|
296 |
-
citation: `string`, citation for the data set
|
297 |
-
url: `string`, url for information about the data set
|
298 |
-
label_classes: `list[string]`, the list of classes if the label is
|
299 |
-
categorical. If not provided, then the label will be of type
|
300 |
-
`datasets.Value('float32')`.
|
301 |
-
multi_label: `boolean`, True if the task is multi-label
|
302 |
-
dev_column: `string`, name for the development subset
|
303 |
-
**kwargs: keyword arguments forwarded to super.
|
304 |
-
"""
|
305 |
-
super(LexGlueConfig, self).__init__(version=datasets.Version("1.0.0", ""), **kwargs)
|
306 |
-
self.text_column = text_column
|
307 |
-
self.label_column = label_column
|
308 |
-
self.label_classes = label_classes
|
309 |
-
self.multi_label = multi_label
|
310 |
-
self.dev_column = dev_column
|
311 |
-
self.url = url
|
312 |
-
self.data_url = data_url
|
313 |
-
self.data_file = data_file
|
314 |
-
self.citation = citation
|
315 |
-
|
316 |
-
|
317 |
-
class LexGLUE(datasets.GeneratorBasedBuilder):
|
318 |
-
"""LexGLUE: A Benchmark Dataset for Legal Language Understanding in English. Version 1.0"""
|
319 |
-
|
320 |
-
BUILDER_CONFIGS = [
|
321 |
-
LexGlueConfig(
|
322 |
-
name="ecthr_a",
|
323 |
-
description=textwrap.dedent(
|
324 |
-
"""\
|
325 |
-
The European Court of Human Rights (ECtHR) hears allegations that a state has
|
326 |
-
breached human rights provisions of the European Convention of Human Rights (ECHR).
|
327 |
-
For each case, the dataset provides a list of factual paragraphs (facts) from the case description.
|
328 |
-
Each case is mapped to articles of the ECHR that were violated (if any)."""
|
329 |
-
),
|
330 |
-
text_column="facts",
|
331 |
-
label_column="violated_articles",
|
332 |
-
label_classes=ECTHR_ARTICLES,
|
333 |
-
multi_label=True,
|
334 |
-
dev_column="dev",
|
335 |
-
data_url="https://zenodo.org/record/5532997/files/ecthr.tar.gz",
|
336 |
-
data_file="ecthr.jsonl",
|
337 |
-
url="https://archive.org/details/ECtHR-NAACL2021",
|
338 |
-
citation=textwrap.dedent(
|
339 |
-
"""\
|
340 |
-
@inproceedings{chalkidis-etal-2021-paragraph,
|
341 |
-
title = "Paragraph-level Rationale Extraction through Regularization: A case study on {E}uropean Court of Human Rights Cases",
|
342 |
-
author = "Chalkidis, Ilias and
|
343 |
-
Fergadiotis, Manos and
|
344 |
-
Tsarapatsanis, Dimitrios and
|
345 |
-
Aletras, Nikolaos and
|
346 |
-
Androutsopoulos, Ion and
|
347 |
-
Malakasiotis, Prodromos",
|
348 |
-
booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
|
349 |
-
month = jun,
|
350 |
-
year = "2021",
|
351 |
-
address = "Online",
|
352 |
-
publisher = "Association for Computational Linguistics",
|
353 |
-
url = "https://aclanthology.org/2021.naacl-main.22",
|
354 |
-
doi = "10.18653/v1/2021.naacl-main.22",
|
355 |
-
pages = "226--241",
|
356 |
-
}
|
357 |
-
}"""
|
358 |
-
),
|
359 |
-
),
|
360 |
-
LexGlueConfig(
|
361 |
-
name="ecthr_b",
|
362 |
-
description=textwrap.dedent(
|
363 |
-
"""\
|
364 |
-
The European Court of Human Rights (ECtHR) hears allegations that a state has
|
365 |
-
breached human rights provisions of the European Convention of Human Rights (ECHR).
|
366 |
-
For each case, the dataset provides a list of factual paragraphs (facts) from the case description.
|
367 |
-
Each case is mapped to articles of ECHR that were allegedly violated (considered by the court)."""
|
368 |
-
),
|
369 |
-
text_column="facts",
|
370 |
-
label_column="allegedly_violated_articles",
|
371 |
-
label_classes=ECTHR_ARTICLES,
|
372 |
-
multi_label=True,
|
373 |
-
dev_column="dev",
|
374 |
-
url="https://archive.org/details/ECtHR-NAACL2021",
|
375 |
-
data_url="https://zenodo.org/record/5532997/files/ecthr.tar.gz",
|
376 |
-
data_file="ecthr.jsonl",
|
377 |
-
citation=textwrap.dedent(
|
378 |
-
"""\
|
379 |
-
@inproceedings{chalkidis-etal-2021-paragraph,
|
380 |
-
title = "Paragraph-level Rationale Extraction through Regularization: A case study on {E}uropean Court of Human Rights Cases",
|
381 |
-
author = "Chalkidis, Ilias
|
382 |
-
and Fergadiotis, Manos
|
383 |
-
and Tsarapatsanis, Dimitrios
|
384 |
-
and Aletras, Nikolaos
|
385 |
-
and Androutsopoulos, Ion
|
386 |
-
and Malakasiotis, Prodromos",
|
387 |
-
booktitle = "Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies",
|
388 |
-
year = "2021",
|
389 |
-
address = "Online",
|
390 |
-
url = "https://aclanthology.org/2021.naacl-main.22",
|
391 |
-
}
|
392 |
-
}"""
|
393 |
-
),
|
394 |
-
),
|
395 |
-
LexGlueConfig(
|
396 |
-
name="eurlex",
|
397 |
-
description=textwrap.dedent(
|
398 |
-
"""\
|
399 |
-
European Union (EU) legislation is published in EUR-Lex portal.
|
400 |
-
All EU laws are annotated by EU's Publications Office with multiple concepts from the EuroVoc thesaurus,
|
401 |
-
a multilingual thesaurus maintained by the Publications Office.
|
402 |
-
The current version of EuroVoc contains more than 7k concepts referring to various activities
|
403 |
-
of the EU and its Member States (e.g., economics, health-care, trade).
|
404 |
-
Given a document, the task is to predict its EuroVoc labels (concepts)."""
|
405 |
-
),
|
406 |
-
text_column="text",
|
407 |
-
label_column="labels",
|
408 |
-
label_classes=EUROVOC_CONCEPTS,
|
409 |
-
multi_label=True,
|
410 |
-
dev_column="dev",
|
411 |
-
url="https://zenodo.org/record/5363165#.YVJOAi8RqaA",
|
412 |
-
data_url="https://zenodo.org/record/5532997/files/eurlex.tar.gz",
|
413 |
-
data_file="eurlex.jsonl",
|
414 |
-
citation=textwrap.dedent(
|
415 |
-
"""\
|
416 |
-
@inproceedings{chalkidis-etal-2021-multieurlex,
|
417 |
-
author = {Chalkidis, Ilias and
|
418 |
-
Fergadiotis, Manos and
|
419 |
-
Androutsopoulos, Ion},
|
420 |
-
title = {MultiEURLEX -- A multi-lingual and multi-label legal document
|
421 |
-
classification dataset for zero-shot cross-lingual transfer},
|
422 |
-
booktitle = {Proceedings of the 2021 Conference on Empirical Methods
|
423 |
-
in Natural Language Processing},
|
424 |
-
year = {2021},
|
425 |
-
location = {Punta Cana, Dominican Republic},
|
426 |
-
}
|
427 |
-
}"""
|
428 |
-
),
|
429 |
-
),
|
430 |
-
LexGlueConfig(
|
431 |
-
name="scotus",
|
432 |
-
description=textwrap.dedent(
|
433 |
-
"""\
|
434 |
-
The US Supreme Court (SCOTUS) is the highest federal court in the United States of America
|
435 |
-
and generally hears only the most controversial or otherwise complex cases which have not
|
436 |
-
been sufficiently well solved by lower courts. This is a single-label multi-class classification
|
437 |
-
task, where given a document (court opinion), the task is to predict the relevant issue areas.
|
438 |
-
The 14 issue areas cluster 278 issues whose focus is on the subject matter of the controversy (dispute)."""
|
439 |
-
),
|
440 |
-
text_column="text",
|
441 |
-
label_column="issueArea",
|
442 |
-
label_classes=SCDB_ISSUE_AREAS,
|
443 |
-
multi_label=False,
|
444 |
-
dev_column="dev",
|
445 |
-
url="http://scdb.wustl.edu/data.php",
|
446 |
-
data_url="https://zenodo.org/record/5532997/files/scotus.tar.gz",
|
447 |
-
data_file="scotus.jsonl",
|
448 |
-
citation=textwrap.dedent(
|
449 |
-
"""\
|
450 |
-
@misc{spaeth2020,
|
451 |
-
author = {Harold J. Spaeth and Lee Epstein and Andrew D. Martin, Jeffrey A. Segal
|
452 |
-
and Theodore J. Ruger and Sara C. Benesh},
|
453 |
-
year = {2020},
|
454 |
-
title ={{Supreme Court Database, Version 2020 Release 01}},
|
455 |
-
url= {http://Supremecourtdatabase.org},
|
456 |
-
howpublished={Washington University Law}
|
457 |
-
}"""
|
458 |
-
),
|
459 |
-
),
|
460 |
-
LexGlueConfig(
|
461 |
-
name="ledgar",
|
462 |
-
description=textwrap.dedent(
|
463 |
-
"""\
|
464 |
-
LEDGAR dataset aims contract provision (paragraph) classification.
|
465 |
-
The contract provisions come from contracts obtained from the US Securities and Exchange Commission (SEC)
|
466 |
-
filings, which are publicly available from EDGAR. Each label represents the single main topic
|
467 |
-
(theme) of the corresponding contract provision."""
|
468 |
-
),
|
469 |
-
text_column="text",
|
470 |
-
label_column="clause_type",
|
471 |
-
label_classes=LEDGAR_CATEGORIES,
|
472 |
-
multi_label=False,
|
473 |
-
dev_column="dev",
|
474 |
-
url="https://metatext.io/datasets/ledgar",
|
475 |
-
data_url="https://zenodo.org/record/5532997/files/ledgar.tar.gz",
|
476 |
-
data_file="ledgar.jsonl",
|
477 |
-
citation=textwrap.dedent(
|
478 |
-
"""\
|
479 |
-
@inproceedings{tuggener-etal-2020-ledgar,
|
480 |
-
title = "{LEDGAR}: A Large-Scale Multi-label Corpus for Text Classification of Legal Provisions in Contracts",
|
481 |
-
author = {Tuggener, Don and
|
482 |
-
von D{\"a}niken, Pius and
|
483 |
-
Peetz, Thomas and
|
484 |
-
Cieliebak, Mark},
|
485 |
-
booktitle = "Proceedings of the 12th Language Resources and Evaluation Conference",
|
486 |
-
year = "2020",
|
487 |
-
address = "Marseille, France",
|
488 |
-
url = "https://aclanthology.org/2020.lrec-1.155",
|
489 |
-
}
|
490 |
-
}"""
|
491 |
-
),
|
492 |
-
),
|
493 |
-
LexGlueConfig(
|
494 |
-
name="unfair_tos",
|
495 |
-
description=textwrap.dedent(
|
496 |
-
"""\
|
497 |
-
The UNFAIR-ToS dataset contains 50 Terms of Service (ToS) from on-line platforms (e.g., YouTube,
|
498 |
-
Ebay, Facebook, etc.). The dataset has been annotated on the sentence-level with 8 types of
|
499 |
-
unfair contractual terms (sentences), meaning terms that potentially violate user rights
|
500 |
-
according to the European consumer law."""
|
501 |
-
),
|
502 |
-
text_column="text",
|
503 |
-
label_column="labels",
|
504 |
-
label_classes=UNFAIR_CATEGORIES,
|
505 |
-
multi_label=True,
|
506 |
-
dev_column="val",
|
507 |
-
url="http://claudette.eui.eu",
|
508 |
-
data_url="https://zenodo.org/record/5532997/files/unfair_tos.tar.gz",
|
509 |
-
data_file="unfair_tos.jsonl",
|
510 |
-
citation=textwrap.dedent(
|
511 |
-
"""\
|
512 |
-
@article{lippi-etal-2019-claudette,
|
513 |
-
title = "{CLAUDETTE}: an automated detector of potentially unfair clauses in online terms of service",
|
514 |
-
author = {Lippi, Marco
|
515 |
-
and Pałka, Przemysław
|
516 |
-
and Contissa, Giuseppe
|
517 |
-
and Lagioia, Francesca
|
518 |
-
and Micklitz, Hans-Wolfgang
|
519 |
-
and Sartor, Giovanni
|
520 |
-
and Torroni, Paolo},
|
521 |
-
journal = "Artificial Intelligence and Law",
|
522 |
-
year = "2019",
|
523 |
-
publisher = "Springer",
|
524 |
-
url = "https://doi.org/10.1007/s10506-019-09243-2",
|
525 |
-
pages = "117--139",
|
526 |
-
}"""
|
527 |
-
),
|
528 |
-
),
|
529 |
-
LexGlueConfig(
|
530 |
-
name="case_hold",
|
531 |
-
description=textwrap.dedent(
|
532 |
-
"""\
|
533 |
-
The CaseHOLD (Case Holdings on Legal Decisions) dataset contains approx. 53k multiple choice
|
534 |
-
questions about holdings of US court cases from the Harvard Law Library case law corpus.
|
535 |
-
Holdings are short summaries of legal rulings accompany referenced decisions relevant for the present case.
|
536 |
-
The input consists of an excerpt (or prompt) from a court decision, containing a reference
|
537 |
-
to a particular case, while the holding statement is masked out. The model must identify
|
538 |
-
the correct (masked) holding statement from a selection of five choices."""
|
539 |
-
),
|
540 |
-
text_column="text",
|
541 |
-
label_column="labels",
|
542 |
-
dev_column="dev",
|
543 |
-
multi_label=False,
|
544 |
-
label_classes=CASEHOLD_LABELS,
|
545 |
-
url="https://github.com/reglab/casehold",
|
546 |
-
data_url="https://zenodo.org/record/5532997/files/casehold.tar.gz",
|
547 |
-
data_file="casehold.csv",
|
548 |
-
citation=textwrap.dedent(
|
549 |
-
"""\
|
550 |
-
@inproceedings{Zheng2021,
|
551 |
-
author = {Lucia Zheng and
|
552 |
-
Neel Guha and
|
553 |
-
Brandon R. Anderson and
|
554 |
-
Peter Henderson and
|
555 |
-
Daniel E. Ho},
|
556 |
-
title = {When Does Pretraining Help? Assessing Self-Supervised Learning for
|
557 |
-
Law and the CaseHOLD Dataset},
|
558 |
-
year = {2021},
|
559 |
-
booktitle = {International Conference on Artificial Intelligence and Law},
|
560 |
-
}"""
|
561 |
-
),
|
562 |
-
),
|
563 |
-
]
|
564 |
-
|
565 |
-
def _info(self):
|
566 |
-
if self.config.name == "case_hold":
|
567 |
-
features = {
|
568 |
-
"context": datasets.Value("string"),
|
569 |
-
"endings": datasets.features.Sequence(datasets.Value("string")),
|
570 |
-
}
|
571 |
-
elif "ecthr" in self.config.name:
|
572 |
-
features = {"text": datasets.features.Sequence(datasets.Value("string"))}
|
573 |
-
else:
|
574 |
-
features = {"text": datasets.Value("string")}
|
575 |
-
if self.config.multi_label:
|
576 |
-
features["labels"] = datasets.features.Sequence(datasets.ClassLabel(names=self.config.label_classes))
|
577 |
-
else:
|
578 |
-
features["label"] = datasets.ClassLabel(names=self.config.label_classes)
|
579 |
-
return datasets.DatasetInfo(
|
580 |
-
description=self.config.description,
|
581 |
-
features=datasets.Features(features),
|
582 |
-
homepage=self.config.url,
|
583 |
-
citation=self.config.citation + "\n" + MAIN_CITATION,
|
584 |
-
)
|
585 |
-
|
586 |
-
def _split_generators(self, dl_manager):
|
587 |
-
archive = dl_manager.download(self.config.data_url)
|
588 |
-
return [
|
589 |
-
datasets.SplitGenerator(
|
590 |
-
name=datasets.Split.TRAIN,
|
591 |
-
# These kwargs will be passed to _generate_examples
|
592 |
-
gen_kwargs={
|
593 |
-
"filepath": self.config.data_file,
|
594 |
-
"split": "train",
|
595 |
-
"files": dl_manager.iter_archive(archive),
|
596 |
-
},
|
597 |
-
),
|
598 |
-
datasets.SplitGenerator(
|
599 |
-
name=datasets.Split.TEST,
|
600 |
-
# These kwargs will be passed to _generate_examples
|
601 |
-
gen_kwargs={
|
602 |
-
"filepath": self.config.data_file,
|
603 |
-
"split": "test",
|
604 |
-
"files": dl_manager.iter_archive(archive),
|
605 |
-
},
|
606 |
-
),
|
607 |
-
datasets.SplitGenerator(
|
608 |
-
name=datasets.Split.VALIDATION,
|
609 |
-
# These kwargs will be passed to _generate_examples
|
610 |
-
gen_kwargs={
|
611 |
-
"filepath": self.config.data_file,
|
612 |
-
"split": self.config.dev_column,
|
613 |
-
"files": dl_manager.iter_archive(archive),
|
614 |
-
},
|
615 |
-
),
|
616 |
-
]
|
617 |
-
|
618 |
-
def _generate_examples(self, filepath, split, files):
|
619 |
-
"""This function returns the examples in the raw (text) form."""
|
620 |
-
if self.config.name == "case_hold":
|
621 |
-
if "dummy" in filepath:
|
622 |
-
SPLIT_RANGES = {"train": (1, 3), "dev": (3, 5), "test": (5, 7)}
|
623 |
-
else:
|
624 |
-
SPLIT_RANGES = {"train": (1, 45001), "dev": (45001, 48901), "test": (48901, 52501)}
|
625 |
-
for path, f in files:
|
626 |
-
if path == filepath:
|
627 |
-
f = (line.decode("utf-8") for line in f)
|
628 |
-
for id_, row in enumerate(list(csv.reader(f))[SPLIT_RANGES[split][0] : SPLIT_RANGES[split][1]]):
|
629 |
-
yield id_, {
|
630 |
-
"context": row[1],
|
631 |
-
"endings": [row[2], row[3], row[4], row[5], row[6]],
|
632 |
-
"label": str(row[12]),
|
633 |
-
}
|
634 |
-
break
|
635 |
-
elif self.config.multi_label:
|
636 |
-
for path, f in files:
|
637 |
-
if path == filepath:
|
638 |
-
for id_, row in enumerate(f):
|
639 |
-
data = json.loads(row.decode("utf-8"))
|
640 |
-
labels = sorted(
|
641 |
-
list(set(data[self.config.label_column]).intersection(set(self.config.label_classes)))
|
642 |
-
)
|
643 |
-
if data["data_type"] == split:
|
644 |
-
yield id_, {
|
645 |
-
"text": data[self.config.text_column],
|
646 |
-
"labels": labels,
|
647 |
-
}
|
648 |
-
break
|
649 |
-
else:
|
650 |
-
for path, f in files:
|
651 |
-
if path == filepath:
|
652 |
-
for id_, row in enumerate(f):
|
653 |
-
data = json.loads(row.decode("utf-8"))
|
654 |
-
if data["data_type"] == split:
|
655 |
-
yield id_, {
|
656 |
-
"text": data[self.config.text_column],
|
657 |
-
"label": data[self.config.label_column],
|
658 |
-
}
|
659 |
-
break
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
scotus/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1116b7762ec235264723c6f68509572620b616118db623ec6bbc3671fcbb1b96
|
3 |
+
size 39976265
|
scotus/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a7288e9bd124ec344759f49c855a5079b7d98bed2ee86c51c6b2c65b51c4503
|
3 |
+
size 94360448
|
scotus/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b858a6e7fde006e0dcdf4d1253e8eabfe040a84ad13ef45157f00e9b8d5c8b5
|
3 |
+
size 39074686
|
unfair_tos/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cf3a30be32f5635bc269b12685290a27b144ad837391e776a2070ff0c3a521c4
|
3 |
+
size 146864
|
unfair_tos/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09d6d53aa05a1338d399da97f36cae171ca0a85990402318d134a8c668c86182
|
3 |
+
size 500893
|
unfair_tos/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6429ccec1085241b05083880fb73548361bc8fd05a6764152f9b8e43e4c177e6
|
3 |
+
size 217847
|