|
Query-Based Keyphrase Extraction from Long Documents |
|
Martin Docekal, Pavel Smrz |
|
Brno University of Technology |
|
idocekal@fit.vutbr.cz, smrz@fit.vutbr.cz |
|
Abstract |
|
Transformer-based architectures in natural language process- |
|
ing force input size limits that can be problematic when long |
|
documents need to be processed. This paper overcomes this |
|
issue for keyphrase extraction by chunking the long docu- |
|
ments while keeping a global context as a query defining the |
|
topic for which relevant keyphrases should be extracted. The |
|
developed system employs a pre-trained BERT model and |
|
adapts it to estimate the probability that a given text span |
|
forms a keyphrase. We experimented using various context |
|
sizes on two popular datasets, Inspec and SemEval, and a |
|
large novel dataset. The presented results show that a shorter |
|
context with a query overcomes a longer one without the |
|
query on long documents.1 |
|
Introduction |
|
Keyphrase refers to a short language expression describ- |
|
ing the content of a longer text. Due to their concise form, |
|
keyphrases can be used for a quick familiarization with |
|
a document. They also improve the findability of docu- |
|
ments or passages within them. In the bibliographic records, |
|
keyphrase descriptors enable flexible indexing. |
|
Whether a text span is a keyphrase depends on the context |
|
of that span because a keyphrase for a specific topic may |
|
not be a keyphrase for another topic. The presented work |
|
builds on the idea that the topic can be explicitly given as |
|
an input to the keyphrase extraction algorithm in the form |
|
of a query. We approximate such a query with a document’s |
|
title in our experiments. We also investigate the influence of |
|
context size and document structure on the results. |
|
Related Work |
|
Traditional approaches to keyphrase extraction involve |
|
graph-based methods, such as TextRank (Mihalcea and Ta- |
|
rau 2004) and RAKE (Rose et al. 2010). Recently, many |
|
types of neural networks have been used for the task (Lin |
|
and Wang 2019; Sahrawat et al. 2020). Most of the deep |
|
learning work assumes the existence of a title and an abstract |
|
of the document and extracts keyphrases from them because |
|
Copyright © 2022 by the authors. All rights reserved. |
|
1The code is available at https://github.com/KNOT-FIT- |
|
BUT/QBEK.they struggle with longer inputs such as whole scientific ar- |
|
ticles (Kontoulis, Papagiannopoulou, and Tsoumakas 2021). |
|
Some works try to overcome this limitation by first creating |
|
a document summary and then extracting keyphrases from |
|
it (Kontoulis, Papagiannopoulou, and Tsoumakas 2021; Liu |
|
and Iwaihara 2021). Our research follows an alternative |
|
path, compensating for the limited context by a query spec- |
|
ifying a topic. |
|
Model |
|
First, a document is split into parts (contexts), which are fur- |
|
ther processed independently. Then, the devised model esti- |
|
mates the probability that a given continuous text span forms |
|
a keyphrase. It looks for boundaries tsandte, corresponding |
|
to the text span’s start and end, respectively. The inspiration |
|
for this approach comes from the task of reading compre- |
|
hension, where a similar technique is used to search for po- |
|
tential answers to a question in an input text (Devlin et al. |
|
2019). Formally, the model estimates probability: |
|
P(ts; tejx) =P(tsjx)P(tejx); (1) |
|
where xis the input sequence. It assumes that the start and |
|
the end of a text span < ts; te>are independent. The prob- |
|
abilities P(tsjx)andP(tejx)are obtained in the following |
|
way: |
|
P(tjx) =sigmoid (wT |
|
BERT (x)[] +b); (2) |
|
wherestands for the start and end positions. Weights |
|
wsandweare learnable vectors, bis the scalar bias and |
|
BERT (x)[i]is BERT (Devlin et al. 2019) vector representa- |
|
tion of a token from sequence xon position i. See the model |
|
illustration in Figure 1. |
|
The task of predicting whether a given token is the start |
|
token of a span or the end token could be seen as binary clas- |
|
sification with two classes start/end andnot start/not end , |
|
respectively. The binary cross-entropy ( BCE ) loss function |
|
is used for training in the following way: |
|
BCE( vs; gs) + BCE( ve; ge); (3) |
|
where vsis a vector of probabilities that a token is the start |
|
token of a span, for each token in the input, and veis vec- |
|
tor computed analogously, but for the ends. The gsandge |
|
are vectors of ground truth probabilities of starts or ends, |
|
respectively.x = [CLS] title [SEP] context [SEP] |
|
P(t |x)s |
|
P(t |x)e |
|
BERT |
|
BERT |
|
Linear Layer |
|
SigmoidBERT(x)[i]Figure 1: Illustration of our model with query at the input. |
|
We work with two types of inputs. One consists of a text |
|
fragment such as a sentence, and the other uses a query (doc- |
|
ument title) and a text segment, separated by a special token. |
|
Various context sizes are explored in our experiments. The |
|
context size determines how big is the document part the |
|
model sees at once. Every context part of a document is pro- |
|
cessed independently. The final list of keyphrases is created |
|
by collecting keyphrase spans with their scores and selecting |
|
the top ones. |
|
Datasets |
|
Besides two standard datasets for keyphrase extraction, we |
|
created and used a novel dataset of long documents, referred |
|
to a Library, and we also prepared an unstructured version |
|
of the SemEval-2010 dataset. A comparison of the datasets |
|
is given in Table 1. |
|
dataset train val. testsentences |
|
(train) |
|
SemEval-2010 130 14 100 66 428 (72%) |
|
Unstructured- |
|
SemEval-2010130 14 100 45 346 (67%) |
|
Inspec 1 000 500 500 5 894 (25%) |
|
Library 48 879 499 499 298 217 589 (94%) |
|
Table 1: The number of documents in each split along with |
|
the total number of sentences in a train set. The percentage in |
|
the sentences column is the proportion of sentences without |
|
keyphrases. |
|
We had to annotate the spans that represent given |
|
keyphrases in the text as the discussed datasets provide just a |
|
list of associated keyphrases with no information about their |
|
actual positions. The search was case insensitive and the |
|
Porter stemmer was utilized for the SemEval and Hulth2003 |
|
(Inspec) datasets. For the Library dataset, as it is in Czech, |
|
theMorphoDiTa lemmatizer2was used. |
|
SemEval-2010 (Kim et al. 2010) consists of whole |
|
plain texts from scientific articles. The dataset provides |
|
keyphrases provided by authors and readers. As it is |
|
common practice (Kim et al. 2010; Kontoulis, Papa- |
|
giannopoulou, and Tsoumakas 2021), we use a combina- |
|
tion of both in our experiments. Our validation dataset was |
|
2http://hdl.handle.net/11858/00-097C-0000-0023-43CD-0created by randomly choosing a subset of the train set. As |
|
the original data source does not explicitly provide the ti- |
|
tles, which we need to formulate a query, we have manually |
|
extracted the title of each document from the plain text. |
|
Documents in this dataset have a well-formed structure. |
|
They contain a title and abstract and are divided into sections |
|
introduced with a headline. As we want to investigate the |
|
influence of such structure on results, we have made an un- |
|
structured version of this dataset. We downloaded the orig- |
|
inal PDFs and used the GROBID3to get a structured XML |
|
version of them. We kept only the text from the document’s |
|
main body while the parts such as title, abstract, authors, |
|
or section headlines were removed. Nevertheless, document |
|
keyphrase annotations remain the same. We call this dataset |
|
Unstructured-SemEval-2010. The name SemEval is used to |
|
name these two collectively. |
|
Inspec (Hulth 2003) contains a set of title-abstract pairs |
|
collected from scientific articles. For each abstract, there are |
|
two sets of keyphrases — controlled , which are restricted to |
|
the Inspec thesaurus, and uncontrolled that can contain any |
|
suitable keyphrase. To be comparable with previous works |
|
(Hulth 2003; Liu and Iwaihara 2021), we used only the un- |
|
controlled set in our experiments. |
|
Library is a newly created dataset that takes advantage |
|
of a large body of scanned documents, provided by Czech |
|
libraries, that were converted to text by OCR software. This |
|
way of getting the text is unreliable, so the dataset contains |
|
many errors on the word and character level. The dataset |
|
builds on the documents where the language was recognized |
|
as ‘Czech’ by the OCR software. |
|
All used documents in the original data source are rep- |
|
resented by their significant content (the average number |
|
of characters per document is 529 276) and metadata. The |
|
metadata contains (not for all) keyphrases and document lan- |
|
guage annotations. We did not ask annotators to annotate |
|
each document. Instead, we selected metadata fields used by |
|
librarians as keyphrase annotations. So, our data and meta- |
|
data come from the real-world environment of Czech li- |
|
braries. We have filtered out all documents with less than |
|
five keyphrases. |
|
Documents come from more than 290 categories. Vari- |
|
ous topics such as mathematics, psychology, belles lettres, |
|
music, and computer science are covered. Not all anno- |
|
tated keyphrases can be extracted from the text. Consid- |
|
ering the lemmatization, the test set annotations contain |
|
about 53% of extractive keyphrases. Bibliographic field Title |
|
(MARC 2454) is used as the query. Note that the field may |
|
contain additional information to the title, such as authors. |
|
Experimental Setup |
|
The implemented system builds on PyTorch5and PyTorch |
|
Lightning6tools. The BERT part of the model uses the |
|
3https://github.com/kermitt2/grobid |
|
4https://www.loc.gov/marc/bibliographic/bd245.html |
|
5https://pytorch.org/ |
|
6https://www.pytorchlightning.ai/1359 19 370:120:140:160:180:2 |
|
input size [sentences]F1@10sentences only sentences + query |
|
sentences only (unst) sentences + query (unst) |
|
Figure 2: Results for SemEval-2010 and Unstructured- |
|
SemEval-2010 test set. The light red and blue areas are con- |
|
fidence intervals with a confidence level of 0.95. Each point |
|
corresponds to an average of five runs. |
|
implementation of BERT by Hugging Face7and it is ini- |
|
tialized with pre-trained weights of bert-base-multilingual- |
|
cased . These weights are also optimized during fine-tuning. |
|
The Adam optimiser with weight decay (Loshchilov and |
|
Hutter 2017) is used in all the experiments. The evaluation |
|
during training on the validation set is done every 4 000 op- |
|
timization steps for the Library dataset and every 50 steps |
|
for Inspec (25 for whole abstracts with titles). For SemEval |
|
datasets, the number of steps differs among experiments. |
|
Early stopping with patience 3 is applied, so the training |
|
ends when the model stops improving. Batch size 128 is |
|
used for experiments with the Library dataset, and batch |
|
size 32 is used for Inspec and SemEval datasets. The learn- |
|
ing rate 1E-06 is used for the experiments with SemEval |
|
datasets, while it is set to the value of 1E-05 for all other |
|
datasets. Inputs longer than a maximum input size are split |
|
into sequences of roughly the same size in a way that for- |
|
bids splitting of keyphrase spans. In edge cases (split is not |
|
possible), the input is truncated. No predicted span is longer |
|
than 6 tokens. |
|
The official script for SemEval-2010 is used for evalua- |
|
tion. However, the normalization of keyphrases is different |
|
for the Library dataset as we have used the mentioned Czech |
|
lemmatizer instead of the original stemmer. We use the F1 |
|
over the top five (F1@5) candidates for the Library dataset |
|
and over the top ten (F1@10) for the rest. |
|
Experiments |
|
The performed experiments investigate the influence of |
|
queries on four different datasets, the output quality with |
|
various context sizes, and the impact of the document struc- |
|
ture. |
|
The first set of experiments is performed on long docu- |
|
ments with a well-formed structure from the SemEval-2010 |
|
7https://huggingface.co/1359 19 370255075 |
|
input size [sentences]too long inputs [%]sentences only sentences + query |
|
Figure 3: The proportion of inputs longer than the maximum |
|
input size for SemEval-2010 train set. |
|
1359 19 370255075 |
|
input size [sentences]proportion [%] |
|
Figure 4: The proportion of contexts containing at least one |
|
section headline as a substring for SemEval-2010 test set. |
|
dataset and compares them with SemEval’s unstructured |
|
version. Figure 2 shows that inputs with a query are bet- |
|
ter than those without a query, but the last point. For the |
|
structured input, it can be seen that from the point with 19 |
|
sentences, the performance of input with query stops with |
|
the fast growth. It correlates with Figure 3 showing the sat- |
|
uration of the model input. Notice that from 19 sentences, |
|
the input becomes more saturated, and the splitting strategy |
|
starts shrinking contexts. |
|
It is not surprising that the nominal values are lower for |
|
unstructured inputs. On the other hand, it is clear that the |
|
query has a bigger influence on the unstructured version, es- |
|
pecially for short context sizes, because the average abso- |
|
lute difference among results (with- and without a query) for |
|
each context size is 2.2% compared to 1.29% for the struc- |
|
tured one. |
|
Looking at the curve of results with a query on an un- |
|
structured version, we assume that the model can exploit a |
|
document structure without explicitly tagging it with special |
|
tokens because additional context size above the three sen- |
|
tences is not much beneficial compared to the case with the |
|
document structure. This hypothesis is supported by the fact |
|
that the proportion of the context containing structured infor- |
|
mation grows with context size, as is demonstrated in Figure |
|
4 showing the proportion of contexts containing a section |
|
headline. |
|
The second set of experiments was performed on our Li- |
|
brary dataset. The results can be seen in Figure 5. We have |
|
chosen F1@5 because only approximately half of the doc- |
|
uments have ten and more keyphrases. Again, the results |
|
show that queries are beneficial. Also, it can be seen that |
|
the shape of the query curve is similar to Unstructured- |
|
SemEval-2010. The average absolute difference between the |
|
version with and without query is now 3.1%. For F1@10, it1359 19 370:10:15 |
|
input size [sentences]F1@5sentences only sentences + query |
|
Figure 5: Results for Library test set for various context |
|
sizes. The light red area symbolizes a confidence interval |
|
with a confidence level of 0.95. Each point is average from |
|
three runs. |
|
Inspec SemEval 2010 |
|
F1@10 F1@10 |
|
TextRank 15.28 6.55 |
|
KFBS + BK-Rank 46.62 15.59 |
|
DistilRoBERTa + TF-IDF - 16.2 |
|
context |
|
[sentences]query |
|
whole document 7 39.67 - |
|
17 40.26 14.96 |
|
3 39.95 16.06 |
|
197 - 17.18 |
|
3 - 18.56 |
|
Table 2: Comparison of achieved results with other work. |
|
KFBS + BK-Rank and TextRank is from (Liu and Iwaihara |
|
2021). The DistilRoBERTa + TF-IDF is from (Kontoulis, |
|
Papagiannopoulou, and Tsoumakas 2021). Our results are |
|
averages from five runs. |
|
is 2.3, which is close to the value for the unstructured version |
|
of SemEval. |
|
The last set of experiments is done on the Inspec dataset, |
|
which has only titles and abstracts. The purpose is to in- |
|
vestigate the influence of a query on short inputs containing |
|
mainly salient sentences. Results are summarized in Table 2, |
|
which also compares our results with other works. It shows |
|
that the results for a single sentence, a single sentence with |
|
a title, and whole abstract with a title are similar. The ex- |
|
planation can be that the abstract contains mainly salient |
|
sentences containing keyphrases, and also, the abstract it- |
|
self defines the topic of the article. A similar observation |
|
is presented in (Liu and Iwaihara 2021), where the version |
|
without summarization gives similar results as the extraction |
|
performed on a summary. |
|
Conclusions |
|
We have conducted experiments that show that query-based |
|
keyphrase extraction is promising for processing long doc- |
|
uments. Our experiments show the relationship between |
|
the context size and the performance of the BERT-basedkeyphrase extractor. The developed model was evaluated on |
|
four datasets; one of them is non-English. The datasets al- |
|
lowed us to find when the query-based approach is benefi- |
|
cial and when not. It was shown that a query gives no bene- |
|
fit when extracting keyphrases from abstracts. On the other |
|
hand, it is beneficial for long documents, particularly those |
|
without a well-formed document structure on short context |
|
sizes. |
|
Acknowledgment |
|
This work was supported by the Technology Agency of the |
|
Czech Republic, Grant FW03010656 – MASAPI: Multilin- |
|
gual assistant for searching, analysing and processing infor- |
|
mation and decision support. The computation used the in- |
|
frastructure supported by the Ministry of Education, Youth |
|
and Sports of the Czech Republic through the e-INFRA CZ |
|
(ID:90140). |
|
References |
|
Devlin, J.; Chang, M.-W.; Lee, K.; and Toutanova, K. 2019. |
|
BERT: Pre-training of deep bidirectional transformers for lan- |
|
guage understanding. In Proceedings of the 2019 NAACL Con- |
|
ference: Human Language Technologies, Volume 1 (Long and |
|
Short Papers) , 4171–4186. Minneapolis, Minnesota: Associa- |
|
tion for Computational Linguistics. |
|
Hulth, A. 2003. Improved automatic keyword extraction given |
|
more linguistic knowledge. In Proceedings of the 2003 EMNLP |
|
Conference , 216–223. |
|
Kim, S. N.; Medelyan, O.; Kan, M.-Y .; and Baldwin, T. 2010. |
|
SemEval-2010 task 5 : Automatic keyphrase extraction from |
|
scientific articles. In Proceedings of the 5th SemEval , 21–26. |
|
Uppsala, Sweden: Association for Computational Linguistics. |
|
Kontoulis, C. G.; Papagiannopoulou, E.; and Tsoumakas, G. |
|
2021. Keyphrase extraction from scientific articles via extrac- |
|
tive summarization. In Proceedings of the Second SDP Work- |
|
shop , 49–55. Online: Association for Computational Linguis- |
|
tics. |
|
Lin, Z.-L., and Wang, C.-J. 2019. Keyword extraction |
|
with character-level convolutional neural tensor networks. In |
|
Pacific-Asia Conference on Knowledge Discovery and Data |
|
Mining , 400–413. Springer. |
|
Liu, T., and Iwaihara, M. 2021. Supervised learning of |
|
keyphrase extraction utilizing prior summarization. In Ke, H.- |
|
R.; Lee, C. S.; and Sugiyama, K., eds., Towards Open and |
|
Trustworthy Digital Societies , 157–166. Cham: Springer In- |
|
ternational Publishing. |
|
Loshchilov, I., and Hutter, F. 2017. Decoupled weight decay |
|
regularization. arXiv preprint arXiv:1711.05101 . |
|
Mihalcea, R., and Tarau, P. 2004. Textrank: Bringing order |
|
into text. In Proceedings of the 2004 conference on EMNLP , |
|
404–411. |
|
Rose, S.; Engel, D.; Cramer, N.; and Cowley, W. 2010. Au- |
|
tomatic keyword extraction from individual documents. Text |
|
mining: applications and theory 1:1–20. |
|
Sahrawat, D.; Mahata, D.; Zhang, H.; Kulkarni, M.; Sharma, |
|
A.; Gosangi, R.; Stent, A.; Kumar, Y .; Shah, R. R.; and Zim- |
|
mermann, R. 2020. Keyphrase extraction as sequence labeling |
|
using contextualized embeddings. In ECIR , 328–335. Springer. |