--- license: cc-by-4.0 configs: - config_name: split_0 data_files: - split: train path: split_0/train-* - split: validation path: split_0/validation-* - split: test path: split_0/test-* - config_name: split_1 data_files: - split: train path: split_1/train-* - split: validation path: split_1/validation-* - split: test path: split_1/test-* - config_name: split_2 data_files: - split: train path: split_2/train-* - split: validation path: split_2/validation-* - split: test path: split_2/test-* - config_name: split_3 data_files: - split: train path: split_3/train-* - split: validation path: split_3/validation-* - split: test path: split_3/test-* - config_name: split_4 data_files: - split: train path: split_4/train-* - split: validation path: split_4/validation-* - split: test path: split_4/test-* - config_name: split_5 data_files: - split: train path: split_5/train-* - split: validation path: split_5/validation-* - split: test path: split_5/test-* - config_name: split_6 data_files: - split: train path: split_6/train-* - split: validation path: split_6/validation-* - split: test path: split_6/test-* - config_name: split_7 data_files: - split: train path: split_7/train-* - split: validation path: split_7/validation-* - split: test path: split_7/test-* - config_name: split_8 data_files: - split: train path: split_8/train-* - split: validation path: split_8/validation-* - split: test path: split_8/test-* - config_name: split_9 data_files: - split: train path: split_9/train-* - split: validation path: split_9/validation-* - split: test path: split_9/test-* --- # LitBank - Project: https://github.com/dbamman/litbank - Data source: https://github.com/dbamman/litbank/commit/3e50db0ffc033d7ccbb94f4d88f6b99210328ed8 - Crossval splits source: https://github.com/dbamman/lrec2020-coref/commit/e30de53743d36d1ea2c9e7292c69477fa332713c ## Details Ten configs of the form f"split_{X}" where X is in range(10) ### Features ``` {'coref_chains': List[List[List[int]]] # list of clusters, each cluster is a list of mentions, each mention is a list of [sent_idx, start, end] inclusive 'doc_name': str 'entities': List[List[{'bio_tags': List[str] 'token': str}]], # list of sentences, each sentence is a list of tokens, each token has a list of bio tags and the token 'events': List[List[{'is_event': bool, 'token': str}]], # list of sentences, each sentence is a list of tokens, each token contains is_event and the token 'meta_info': {'author': str, 'date': str, 'gutenberg_id': str, 'title': str}, 'original_text': str, 'quotes': List[{'attribution': str, 'end': {'sent_id': str, 'token_id': str}, 'quotation': str, 'quote_id': str, 'start': {'sent_id': str, 'token_id': str}}], 'sentences': List[List[str]], } ``` ## Citation ``` @inproceedings{bamman-etal-2019-annotated, title = "An annotated dataset of literary entities", author = "Bamman, David and Popat, Sejal and Shen, Sheng", editor = "Burstein, Jill and Doran, Christy and Solorio, Thamar", booktitle = "Proceedings of the 2019 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)", month = jun, year = "2019", address = "Minneapolis, Minnesota", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/N19-1220", doi = "10.18653/v1/N19-1220", pages = "2138--2144", abstract = "We present a new dataset comprised of 210,532 tokens evenly drawn from 100 different English-language literary texts annotated for ACE entity categories (person, location, geo-political entity, facility, organization, and vehicle). These categories include non-named entities (such as {``}the boy{''}, {``}the kitchen{''}) and nested structure (such as [[the cook]{'}s sister]). In contrast to existing datasets built primarily on news (focused on geo-political entities and organizations), literary texts offer strikingly different distributions of entity categories, with much stronger emphasis on people and description of settings. We present empirical results demonstrating the performance of nested entity recognition models in this domain; training natively on in-domain literary data yields an improvement of over 20 absolute points in F-score (from 45.7 to 68.3), and mitigates a disparate impact in performance for male and female entities present in models trained on news data.", } ``` ### Event detection ``` @inproceedings{sims-etal-2019-literary, title = "Literary Event Detection", author = "Sims, Matthew and Park, Jong Ho and Bamman, David", editor = "Korhonen, Anna and Traum, David and M{\`a}rquez, Llu{\'\i}s", booktitle = "Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics", month = jul, year = "2019", address = "Florence, Italy", publisher = "Association for Computational Linguistics", url = "https://aclanthology.org/P19-1353", doi = "10.18653/v1/P19-1353", pages = "3623--3634", abstract = "In this work we present a new dataset of literary events{---}events that are depicted as taking place within the imagined space of a novel. While previous work has focused on event detection in the domain of contemporary news, literature poses a number of complications for existing systems, including complex narration, the depiction of a broad array of mental states, and a strong emphasis on figurative language. We outline the annotation decisions of this new dataset and compare several models for predicting events; the best performing model, a bidirectional LSTM with BERT token representations, achieves an F1 score of 73.9. We then apply this model to a corpus of novels split across two dimensions{---}prestige and popularity{---}and demonstrate that there are statistically significant differences in the distribution of events for prestige.", } ``` ### Coreference ``` @inproceedings{bamman-etal-2020-annotated, title = "An Annotated Dataset of Coreference in {E}nglish Literature", author = "Bamman, David and Lewke, Olivia and Mansoor, Anya", editor = "Calzolari, Nicoletta and B{\'e}chet, Fr{\'e}d{\'e}ric and Blache, Philippe and Choukri, Khalid and Cieri, Christopher and Declerck, Thierry and Goggi, Sara and Isahara, Hitoshi and Maegaard, Bente and Mariani, Joseph and Mazo, H{\'e}l{\`e}ne and Moreno, Asuncion and Odijk, Jan and Piperidis, Stelios", booktitle = "Proceedings of the Twelfth Language Resources and Evaluation Conference", month = may, year = "2020", address = "Marseille, France", publisher = "European Language Resources Association", url = "https://aclanthology.org/2020.lrec-1.6", pages = "44--54", abstract = "We present in this work a new dataset of coreference annotations for works of literature in English, covering 29,103 mentions in 210,532 tokens from 100 works of fiction published between 1719 and 1922. This dataset differs from previous coreference corpora in containing documents whose average length (2,105.3 words) is four times longer than other benchmark datasets (463.7 for OntoNotes), and contains examples of difficult coreference problems common in literature. This dataset allows for an evaluation of cross-domain performance for the task of coreference resolution, and analysis into the characteristics of long-distance within-document coreference.", language = "English", ISBN = "979-10-95546-34-4", } ```