|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
"""OrangeSum dataset""" |
|
|
|
|
|
import datasets |
|
|
|
|
|
_CITATION = """\ |
|
@article{eddine2020barthez, |
|
title={BARThez: a Skilled Pretrained French Sequence-to-Sequence Model}, |
|
author={Eddine, Moussa Kamal and Tixier, Antoine J-P and Vazirgiannis, Michalis}, |
|
journal={arXiv preprint arXiv:2010.12321}, |
|
year={2020} |
|
} |
|
""" |
|
|
|
_DESCRIPTION = """\ |
|
The OrangeSum dataset was inspired by the XSum dataset. It was created by scraping the "Orange Actu" website: https://actu.orange.fr/. Orange S.A. is a large French multinational telecommunications corporation, with 266M customers worldwide. Scraped pages cover almost a decade from Feb 2011 to Sep 2020. They belong to five main categories: France, world, politics, automotive, and society. The society category is itself divided into 8 subcategories: health, environment, people, culture, media, high-tech, unsual ("insolite" in French), and miscellaneous. |
|
|
|
Each article featured a single-sentence title as well as a very brief abstract, both professionally written by the author of the article. These two fields were extracted from each page, thus creating two summarization tasks: OrangeSum Title and OrangeSum Abstract. |
|
""" |
|
|
|
_URL_DATA = { |
|
"abstract": "https://raw.githubusercontent.com/Tixierae/OrangeSum/main/data/docs/splits/abstract.tgz", |
|
"title": "https://raw.githubusercontent.com/Tixierae/OrangeSum/main/data/docs/splits/title.tgz", |
|
} |
|
|
|
_DOCUMENT = "text" |
|
_SUMMARY = "summary" |
|
|
|
|
|
class OrangeSum(datasets.GeneratorBasedBuilder): |
|
"""OrangeSum: a french abstractive summarization dataset""" |
|
|
|
VERSION = datasets.Version("1.1.0") |
|
|
|
BUILDER_CONFIGS = [ |
|
datasets.BuilderConfig(name="abstract", description="Abstracts used as summaries", version=VERSION), |
|
datasets.BuilderConfig(name="title", description="Titles used as summaries", version=VERSION), |
|
] |
|
|
|
def _info(self): |
|
return datasets.DatasetInfo( |
|
description=_DESCRIPTION, |
|
features=datasets.Features( |
|
{ |
|
_DOCUMENT: datasets.Value("string"), |
|
_SUMMARY: datasets.Value("string"), |
|
} |
|
), |
|
supervised_keys=(_DOCUMENT, _SUMMARY), |
|
homepage="https://github.com/Tixierae/OrangeSum/", |
|
citation=_CITATION, |
|
) |
|
|
|
def _split_generators(self, dl_manager): |
|
"""Returns SplitGenerators.""" |
|
archive = dl_manager.download(_URL_DATA[self.config.name]) |
|
|
|
return [ |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TRAIN, |
|
|
|
gen_kwargs={ |
|
"source_files": dl_manager.iter_archive(archive), |
|
"target_files": dl_manager.iter_archive(archive), |
|
"split": "train", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.TEST, |
|
|
|
gen_kwargs={ |
|
"source_files": dl_manager.iter_archive(archive), |
|
"target_files": dl_manager.iter_archive(archive), |
|
"split": "test", |
|
}, |
|
), |
|
datasets.SplitGenerator( |
|
name=datasets.Split.VALIDATION, |
|
|
|
gen_kwargs={ |
|
"source_files": dl_manager.iter_archive(archive), |
|
"target_files": dl_manager.iter_archive(archive), |
|
"split": "valid", |
|
}, |
|
), |
|
] |
|
|
|
def _generate_examples(self, source_files, target_files, split): |
|
"""Yields examples.""" |
|
expected_source_path = f"{self.config.name}/{split}.source" |
|
expected_target_path = f"{self.config.name}/{split}.target" |
|
for source_path, f_source in source_files: |
|
if source_path == expected_source_path: |
|
for target_path, f_target in target_files: |
|
if target_path == expected_target_path: |
|
for idx, (document, summary) in enumerate(zip(f_source, f_target)): |
|
yield idx, {_DOCUMENT: document.decode("utf-8"), _SUMMARY: summary.decode("utf-8")} |
|
break |
|
break |
|
|