|
Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
KELM: K NOWLEDGE ENHANCED PRE-TRAINED LAN- |
|
GUAGE REPRESENTATIONS WITH MESSAGE PASSING |
|
ONHIERARCHICAL RELATIONAL GRAPHS |
|
Yinquan Lu1,5Haonan Lu2yGuirong Fu3Qun Liu4 |
|
Huawei Technologies Co., Ltd.1OPPO Guangdong Mobile Telecommunications Co., Ltd.2 |
|
ByteDance3Huawei Noah’s Ark Lab4Shanghai AI Laboratory5 |
|
[email protected], [email protected] |
|
[email protected], [email protected] |
|
ABSTRACT |
|
Incorporating factual knowledge into pre-trained language models (PLM) such as |
|
BERT is an emerging trend in recent NLP studies. However, most of the existing |
|
methods combine the external knowledge integration module with a modified |
|
pre-training loss and re-implement the pre-training process on the large-scale |
|
corpus. Re-pretraining these models is usually resource-consuming, and difficult |
|
to adapt to another domain with a different knowledge graph (KG). Besides, |
|
those works either cannot embed knowledge context dynamically according to |
|
textual context or struggle with the knowledge ambiguity issue. In this paper, we |
|
propose a novel knowledge-aware language model framework based on fine-tuning |
|
process, which equips PLM with a unified knowledge-enhanced text graph that |
|
contains both text and multi-relational sub-graphs extracted from KG. We design |
|
a hierarchical relational-graph-based message passing mechanism, which allows |
|
the representations of injected KG and text to mutually update each other and can |
|
dynamically select ambiguous mentioned entities that share the same text1. Our |
|
empirical results show that our model can efficiently incorporate world knowledge |
|
from KGs into existing language models such as BERT, and achieve significant |
|
improvement on the machine reading comprehension (MRC) tasks compared with |
|
other knowledge-enhanced models. |
|
1 I NTRODUCTION |
|
Pre-trained language models benefit from the large-scale corpus and can learn complex linguistic |
|
representation (Devlin et al., 2019; Liu et al., 2019b; Yang et al., 2020). Although they have achieved |
|
promising results in many NLP tasks, they neglect to incorporate structured knowledge for language |
|
understanding. Limited by implicit knowledge representation, existing PLMs are still difficult to |
|
learn world knowledge efficiently (Poerner et al., 2019; Yu et al., 2020). For example, hundreds |
|
of related training samples in the corpus are required to understand the fact “ banmeans an official |
|
prohibition or edict against something” for PLMs. |
|
By contrast, knowledge graphs (KGs) explicitly organize the above fact as a triplet “(ban, hypernyms, |
|
prohibition)” . Although domain knowledge can be represented more efficiently in KG form, entities |
|
with different meanings share the same text may happen in a KG (knowledge ambiguity issue). For |
|
example, one can also find “(ban, hypernyms, moldovan monetary unit)” in WordNet (Miller, 1995). |
|
Recently, many efforts have been made on leveraging heterogeneous factual knowledge in KGs to |
|
enhance PLM representations. These models generally adopt two methods: (1). Injecting pre-trained |
|
entity embeddings into PLM explicitly, such as ERNIE (Zhang et al., 2019), which injects entity |
|
embeddings pre-trained on a knowledge graph by using TransE (Bordes et al., 2013). (2). Implicitly |
|
This work is done when Yinquan Lu, Haonan Lu and Guirong Fu work at Huawei Technologies Co., Ltd. |
|
yCorresponding author |
|
1Words or phrases in the text corresponding to certain entities in KGs are often named “entity mentions” . |
|
While entities in KGs that correspond to entity mentions in the text are often named “mentioned entities” |
|
1arXiv:2109.04223v2 [cs.CL] 5 May 2022Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
learning factual knowledge by adding extra pre-training tasks such as entity-level mask, entity-based |
|
replacement prediction, etc. (Wang et al., 2020c; Sun et al., 2020). Some studies use both of the |
|
above two methods such as CokeBERT (Su et al., 2020). |
|
However, as summarized in Table 4 of Appendix, most of the existing knowledge-enhanced PLMs |
|
need to re-pretrain the models based on an additional large-scale corpus, they mainly encounter |
|
two problems below: (1) Incorporating external knowledge during pretraining is usually resource- |
|
consuming and difficult to adapt to other domains with different KGs. By checking the third column |
|
of Table 4 in Appendix, one can see that most of the pretrain-based models use Wiki-related KG as |
|
their injected knowledge source. These models also use English Wikipedia as pre-training corpus. |
|
They either use an additional entity linking tool (e.g. TAGME (Ferragina & Scaiella, 2010)) to align |
|
the entity mention in the text to a single mentioned entity in a Wiki-related KG uniquely or directly |
|
treat hyperlinks in Wikipedia as entity annotations. These models depend heavily on the one-to-one |
|
mapping relationship between Wikipedia corpus and Wiki-related KG, thus they never consider |
|
handling knowledge ambiguity issue. (2) These models with explicit knowledge injection usually use |
|
algorithms like BILINEAR (Yang et al., 2015) to obtain pre-trained KG embeddings, which contain |
|
information about graph structure. Unfortunately, their knowledge context is usually static and cannot |
|
be embedded dynamically according to textual context. |
|
Several works (Qiu et al., 2019; Yang et al., 2019) concentrate on injecting external knowledge based |
|
on fine-tuning PLM on downstream tasks, which is much easier to change the injected KGs and adapt |
|
to relevant domain tasks. They either cannot consider multi-hop relational information, or struggle |
|
with knowledge ambiguity issue. How to fuse heterogeneous information dynamically based on the |
|
fine-tuning process on the downstream tasks and use the information of injected KGs more efficiently |
|
remains a challenge. |
|
Figure 1: Unified Knowledge-enhanced Text Graph (UKET) |
|
consists of three parts corresponding to our model: (1) KG |
|
only part, (2) Entity link to token graph, (3) Text only graph.To overcome the challenges mentioned |
|
above, we propose a novel frame- |
|
work named KELM , which injects |
|
world knowledge from KGs during |
|
the fine-tuning phase by building a |
|
Unified Knowledge-enhanced Text |
|
Graph (UKET) that contains both in- |
|
jected sub-graphs from external knowl- |
|
edge and text. The method extends |
|
the input sentence by extracting sub- |
|
graphs centered on every mentioned |
|
entity from KGs. In this way, we can |
|
get a Unified Knowledge-enhanced |
|
Text Graph as shown in Fig. 1, which is |
|
made of three kinds of graph: (1) The |
|
injected knowledge graphs, referred to |
|
as the “ KG only ” part; (2) The graph |
|
about entity mentions in the text and mentioned entities in KGs, referred to as the “ entity link to |
|
token ” part. Entity mentions in the text are linked with mentioned entities in KGs by string matching, |
|
so one entity mention may trigger several mentioned entities that share the same text in the injected |
|
KGs (e.g. “Ford” in Fig. 1); (3) The “ text only ” part, where the input text sequence is treated as a |
|
fully-connected word graph just like classical Transformer architecture (Vaswani et al., 2017). |
|
Based on this unified graph, we design a novel Hierarchical relational-graph-based Message Passing |
|
(HMP) mechanism to fuse heterogeneous information on the output layer of PLM. The implemen- |
|
tation of HMP is via a Hierarchical Knowledge Enhancement Module as depicted in Fig. 2, which |
|
also consists of three parts, and each part is designed for solving the different problems above: (1) |
|
For reserving the structure information and dynamically embedding injected knowledge, we utilize |
|
a relational GNN (e.g. rGCN (Schlichtkrull et al., 2017)) to aggregate and update representations |
|
of extracted sub-graphs for each injected KG (corresponding to the “KG only” part of UKET). All |
|
mentioned entities and their K-hop neighbors in sub-graphs are initialized by pre-trained vectors |
|
obtained from the classical knowledge graph embedding (KGE) method (we adopt BILINEAR here). |
|
In this way, knowledge context can be dynamically embedded, the structural information about the |
|
graph is also kept; (2) For handling knowledge ambiguity issue and selecting relevant mentioned |
|
entities according to the input context, we leverage a specially designed attention mechanism to |
|
2Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
Figure 2: Framework of KELM (left) and illustrates how to generate knowledge-enriched token |
|
embeddings (right). |
|
weight these ambiguous mentioned entities by using the textual representations of words/tokens to |
|
query the representations of their related mentioned entities in KGs (corresponding to the “entity |
|
link to token” graph of UKET). The attention score can help to select knowledge according to the |
|
input sentence dynamically. By concatenating the outputs of this step with the original outputs |
|
of PLM, we can get a knowledge-enriched representation for each token; (3) For further interac- |
|
tions between knowledge-enriched tokens, we employ a self-attention mechanism that operates on |
|
the fully-connected word graph (corresponding to the “text only” graph of UKET) to allow the |
|
knowledge-enriched representation of each token to further interact with others. |
|
We conduct experiments on the MRC task, which requires a system to comprehend a given text |
|
and answer questions about it. In this paper, to prove the generalization ability of our method, |
|
we evaluate KELM on both the extractive-style MRC task (answers can be found in a span of the |
|
given text) and the multiple-response-items-style MRC task (each question is associated with several |
|
choices for answer-options, the number of correct answer-options is not pre-specified). MRC is a |
|
challenging task and represents a valuable path towards natural language understanding (NLU). With |
|
the rapid increment of knowledge, NLU becomes more difficult since the system needs to absorb new |
|
knowledge continuously. Pre-training models on large-scale corpus is inefficient. Therefore, fine- |
|
tuning the knowledge-enhanced PLM on the downstream tasks directly is crucial in the application.2 |
|
2 R ELATED WORK |
|
2.1 K NOWLEDGE GRAPH EMBEDDING |
|
We denote a directed knowledge graph as G(E;R), whereEandRare sets of entities and relations, |
|
respectively. We also define Fas a set of facts, a fact stored in a KG can be expressed as a triplet |
|
(h;r;t )2F, which indicates a relation rpointing from the head entity hto tail entity t, where |
|
h;t2E andr2R . KGE aims to extract topological information in KG and to learn a set of |
|
low-dimensional representations of entities and relations by knowledge graph completion task (Yang |
|
et al., 2015; Lu & Hu, 2020). |
|
2.2 M ULTI -RELATIONAL GRAPH NEURAL NETWORK |
|
Real-world KGs usually include several relations. However, traditional GNN models such as |
|
GCN (Kipf & Welling, 2017), and GAT (Veli ˇckovi ´c et al., 2018) can only be used in the graph |
|
2Code is available at here. |
|
3Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
with one type of relation. (Schlichtkrull et al., 2017; Haonan et al., 2019) generalizes traditional |
|
GNN models by performing relation-specific aggregation, making it possible to encode relational |
|
graphs. The use of multi-relational GNN makes it possible to encode injected knowledge embeddings |
|
dynamically in SKG and CokeBERT. |
|
2.3 J OINT LANGUAGE AND KNOWLEDGE MODELS |
|
Since BERT was published in 2018, many efforts have been made for further optimization, basically |
|
focusing on the design of the pre-training process and the variation of the encoder. For studies of |
|
knowledge-enhanced PLMs, they also fall into the above two categories or combine both of them |
|
sometimes. Despite their success in leveraging external factual knowledge, the gains are limited by |
|
computing resources, knowledge ambiguity issue, and the expressivity of their methods for the fusion |
|
of heterogeneous information, as summarized in Table 4 of Appendix and the introduction part. |
|
Recent studies notice that the architecture of Transformer treats input sequences as fully-connected |
|
word graphs, thus some of them try to integrate injected KGs and textual context into a unified |
|
data structure. Here we argue that UKET in our KELM is different from the WK graph proposed |
|
in CoLAKE/K-BERT. These two studies heuristically convert textual context and entity-related |
|
sub-graph into input sequences, both entities and relations are treated as input words of the PLM, |
|
then they leverage a Transformer with a masked attention mechanism to encode those sequences |
|
from the embedding layer and pre-train the model based on the large-scale corpus. Unfortunately, it |
|
is not trivial for them to convert the second or higher order neighbors related to textual context (Su |
|
et al., 2020), the structural information about the graph is lost. UKET differs from the WK graph |
|
of CoLAKE/K-BERT in that, instead of converting mentioned entities, relations, and text into a |
|
sequence of words and feeding them together into the input layer of PLM (they unify text and KG into |
|
a sequence), UKET unifies text and KG into a graph. Besides, by using our UKET framework, the |
|
knowledge fusion process of KELM is based on the representation of the last hidden layer of PLM, |
|
making it possible to directly fine-tune the PLM on the downstream tasks without re-pretraining the |
|
model. SKG also utilizes relational GNN to fuse information of KGs and text representation encoded |
|
by PLM. However, SKG only uses GNN to dynamically encode the injected KGs, which corresponds |
|
to part one of Fig. 1. Outputs of SKG are made by directly concatenating outputs of graph encoder |
|
with the outputs of PLM. It cannot select ambiguous knowledge and forbids the interactions between |
|
knowledge-enriched tokens corresponding to part two and part three of Fig. 1, respectively. KT-NET |
|
uses a specially designed attention mechanism to select relevant knowledge from KGs. For example, |
|
it treats all synsets of entity mentions within the WN183as candidate KB concepts. This limits the |
|
ability of KT-NET to select the most relevant mentioned entities4. Moreover, the representations |
|
of injected knowledge are static in KT-NET, they cannot dynamically change according to textual |
|
context, the information about the original graph structure in KG is also lost. |
|
3 M ETHODOLOGY |
|
The architecture of KELM is shown in Fig. 2. It consists of three main modules: (1) PLM Encoding |
|
Module; (2) Hierarchical Knowledge Enhancement Module; (3) Output Module. |
|
3.1 PLM E NCODING MODULE |
|
This module utilizes PLM (e.g.BERT) to encode text to get textual representations for pas- |
|
sages and questions. An input example of the MRC task includes a paragraph and a ques- |
|
tion with a candidate answer, represented as a single sequence of tokens of the length n: |
|
T=f[CLS];Q;(A);[SEP ];P;[SEP ]g=ftign |
|
i=1, whereQ,AandPrepresent all tokens for ques- |
|
tion, candidate answer and paragraph, respectively5.[SEP ]and[CLS]are special tokens in BERT |
|
and defined as a sentence separator and a classification token, respectively. i-th token in the sequence |
|
is represented by ~ht |
|
i2Rdt, wheredtis the last hidden layer size of used PLM. |
|
3A subset of WordNet. |
|
4Refer the example given in the case study of KT-NET, the most relevant concept for the word “ban” is |
|
“forbidding_NN_1” (with the probability of 86.1%), but not “ban_NN_4”. |
|
5Depending on the type of MRC task (extractive-style v.s. multiple-response-items-style), candidate answer |
|
A is not required in the sequence of tokens for the extractive-style MRC task. |
|
4Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
3.2 H IERARCHICAL KNOWLEDGE ENHANCEMENT MODULE |
|
This module is the implementation of our proposed HMP mechanism to fuse information of textual |
|
and graph context. We will formally introduce graph construction for UKET, and the three sub- |
|
processes of HMP in detail in the following sections. |
|
3.2.1 C ONSTRUCTION OF UKET G RAPH |
|
(1) Given a set with jQjelements:fGq |
|
k(Eq |
|
k;Rq |
|
k)gjQj |
|
q=1and input text, where jQjis the total number of |
|
injected KGs, and qindicates the q-th KG. We denote the set of entity mentions related to the q-th |
|
KG asXq=fxq |
|
igjXqj |
|
i=1, wherejXqjis the number of entity mentions in the text. The corresponding |
|
mentioned entities are shared by all tokens in the same entity mention. All mentioned entities |
|
Mq=fmq |
|
igjMqj |
|
i=1are linked with their relevant entity mentions in the text, where jMqjis the number |
|
of mentioned entities in the q-th KG. We define this "entity link to token graph" in Fig. 1 as |
|
Gq |
|
m(Eq |
|
m;Rq |
|
m), whereEq |
|
m=Xq[Mqis the union of entity mentions and their relevant mentioned |
|
entities,Rq |
|
mis a set with only one element that links mentioned entities and their relevant entity |
|
mentions. (2) For i-th mentioned entity mq |
|
iinMq, we retrieve all its K-hop neighbors fNx |
|
mq |
|
igK |
|
x=0 |
|
from theq-th knowledge graph, where Nx |
|
mq |
|
iis a set ofi-th mentioned entity’s x-hop neighbors, hence |
|
we haveN0 |
|
mq |
|
i=fmq |
|
ig. We define "KG-only graph": Gq |
|
s(Eq |
|
s;Rq |
|
s), whereEq |
|
s=SjMqj |
|
i=0SK |
|
x=0Nx |
|
mq |
|
iis the |
|
union of all mentioned entities and their neighbors within the K-hops sub-graph, and Rq |
|
sis a set |
|
of all relations in the extracted sub-graph of q-th KG. (3) The text sequence can be considered as |
|
a fully-connected word graph as pointed out previously. This “text-only graph” can be denoted as |
|
Gt(Et;Rt), whereEtis all tokens in text and Rtis a set with only one element that connects all |
|
tokens. Finally, we define the full hierarchical graph consisting of all three parts fGq |
|
sgjQj |
|
q=1,fGq |
|
mgjQj |
|
q=1, |
|
andGt, as Unified Knowledge-enhanced Text Graph (UKET). |
|
3.2.2 D YNAMICALLY EMBEDDING KNOWLEDGE CONTEXT |
|
We use pre-trained vectors obtained from the KGE method to initialize representations of entities in |
|
Gq |
|
s(Eq |
|
s;Rq |
|
s). Considering the structural information of injected knowledge graph forgotten during |
|
training, we utilize jQjindependent GNN encoders (i.e. g1(:),g2(:)in Fig. 2, which is the case of |
|
injecting two independent KGs in our experiment setting) to dynamically update entity embeddings |
|
ofjQjinjected KGs. We use rGCN to model the multi-relational nature of the knowledge graph. To |
|
updatei-th node ofq-th KG inl-th rGCN layer: |
|
~ sq(l+1) |
|
i =(X |
|
r2Rq |
|
sX |
|
j2Nr |
|
i1 |
|
jNr |
|
ijWq(l) |
|
r~ sq(l) |
|
j)(1) |
|
WhereNr |
|
iis a set of neighbors of i-th node under relation r2Rq |
|
s.Wq(l) |
|
ris trainable weight matrix |
|
atl-th layer and ~ sq(l+1) |
|
i is the hidden state of i-th node at ( l+1)-th layer. After Lupdates,jQjsets |
|
of node embeddings are obtained. The output of the q-th KG can be represented as Sq2RjEq |
|
sjdq, |
|
wherejEq |
|
sjanddqare the numbers of nodes of extracted sub-graph and the dimension of pre-trained |
|
KGE, respectively. |
|
3.2.3 D YNAMICALLY SELECTING SEMANTICS -RELATED MENTIONED ENTITIES |
|
To handle the knowledge ambiguity issue, we introduce an attention layer to weight these ambiguous |
|
mentioned entities by using the textual representations of tokens (outputs of Section 3.1) to query |
|
their semantics-related mentioned entities representations in KGs. Here, we follow the attention |
|
mechanism of GAT to update each entity mention embedding in Gq |
|
m: |
|
~ xq |
|
i=(X |
|
j2Nq |
|
iq |
|
ijWq~ sq |
|
j)(2) |
|
Where~ sq |
|
jis the output embeddings from the q-th rGCN in the previous step. ~ xq |
|
iis the hidden state |
|
ofi-th entity mention xq |
|
iinXq, andNq |
|
iis a set of neighbors of xq |
|
iinGq |
|
m.Wq2Rdoutdinis a |
|
5Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
trainable weight matrix, we set din=dout=dq(thus~ xq |
|
i2Rdq).is a nonlinear activation function. |
|
q |
|
ijis the attention score that weights ambiguous mentioned entities in the q-th KG: |
|
q |
|
ij=exp(LeakyReLU (~ T |
|
q[Wq~ht0 |
|
ijjWq~ sq |
|
j])) |
|
P |
|
k2Nq |
|
iexp(LeakyReLU (~ Tq[Wq~ht0 |
|
ijjWq~ sq |
|
k]))(3) |
|
The representation ~ht |
|
iwith a dimension of dtis projected to the dimension of dq, before using it |
|
to query the related mentioned entity embeddings of Sq:~ht0 |
|
i=Wq |
|
proj~ht |
|
i, whereWq |
|
proj2Rdqdt. |
|
~ q2R2dqis a trainable weight vector. Tis the transposition operation and jjis the concatenation |
|
operation. |
|
Finally, we concatenate outputs of jQjKGs with textual context representation to get final knowledge- |
|
enriched representation: |
|
~hk |
|
i= [~ht |
|
i;~ x1 |
|
i;:::;~ xjQj |
|
i]2Rdt+d1++djQj (4) |
|
If tokentican’t match any entity in q-th KG (say ti=2Xq), we fill~ xq |
|
iin Eq.4 with zeros. Note |
|
that mentioned entities in KGs are not always useful, to prevent noise, we follow (Yang & Mitchell, |
|
2017)’s work and add an extra sentinel node linked to each entity mention in Gq |
|
m. The sentinel node |
|
is initialized by zeros and not trainable, which is the same as the case of no retrieved entities in the |
|
KG. In this way, according to the textual context, KELM can dynamically select mentioned entities |
|
and avoid introducing knowledge noise. |
|
3.2.4 I NTERACTION BETWEEN KNOWLEDGE -ENRICHED TOKEN EMBEDDINGS |
|
To allow knowledge-enriched tokens’ representations to propagate to each other in the text, we |
|
use a fully-connected word graph Gt, with knowledge-enriched representations from outputs of |
|
the previous step, and employ the self-attention mechanism similar to KT-NET to update token’s |
|
embedding. The final representation for i-th token in the text is ~hf |
|
i2R6(dt+d1++djQj). |
|
3.3 O UTPUT MODULE |
|
3.3.1 E XTRACTIVE -STYLE MRC TASK |
|
A simple linear transformation layer and softmax operation are used to predict start and end positions |
|
of answers. For i-th token, the probabilities to be the start and end position of answer span are: |
|
ps |
|
i=exp(wT |
|
s~hf |
|
i) |
|
nP |
|
j=1exp(wTs~hf |
|
j);pe |
|
i=exp(wT |
|
e~hf |
|
i) |
|
nP |
|
j=1exp(wTe~hf |
|
j), wherews;we2R6(dt+d1++djQj)are trainable vectors |
|
andnis the number of tokens. The training loss is calculated by the log-likelihood of the true start |
|
and end positions: L= 1 |
|
NNP |
|
i=1(logps |
|
ys |
|
i+logpe |
|
ye |
|
i), whereNis the total number of examples in the |
|
dataset,ys |
|
iandye |
|
iare the true start and end positions of i-th query’s answer, respectively. During |
|
inference, we pick the span (a;b)with maximum ps |
|
ape |
|
bwhereabas predicted anwser. |
|
3.3.2 M ULTIPLE -RESPONSE -ITEMS -STYLE MRC TASK |
|
Since answers to a given question are independent of each other, to predict the correct probability |
|
of each answer, a fully connected layer followed by a sigmoid function is applied on the final |
|
representation of [CLS]token in BERT. |
|
4 E XPERIMENTS |
|
4.1 D ATASETS |
|
In this paper, we empirically evaluate KELM on both two types of MRC benchmarks in Super- |
|
GLUE (Wang et al., 2020a): ReCoRD (Zhang et al., 2018) (extractive-style) and MultiRC (Khashabi |
|
6Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
et al., 2018) (multiple-response-items-style). Detailed descriptions of the two datasets can be found |
|
in Appendix B. On both datasets, the test set is not public, one has to submit the predicted results to |
|
the organization to get the final test score. Since frequent submissions to probe the unseen test set are |
|
not encouraged, we only submit our best model once for each of the datasets, thus the statistics of the |
|
results (e.g., mean, variance, etc.) are not applicable. We use Exact Match (EM) and (macro-averaged) |
|
F1 as the evaluation metrics. |
|
External Knowledge We adopt knowledge sources the same as used in KT-NET: WordNet and |
|
NELL (Carlson et al., 2010). Representations of injected knowledge are initialized by resources |
|
provided by (Yang & Mitchell, 2017). The size of these embeddings is 100. We retrieve related |
|
knowledge from the two KGs in a given sentence and construct UKET graph (as shown in Section |
|
3.2.1). More details about entity embedding and concepts retrieval are available in Appendix B. |
|
4.2 E XPERIMENTAL SETUPS |
|
Baselines and Comparison Setting Because we use BERT largeas the base model in our method, |
|
we use it as our primary baseline for all tasks. For fair comparison, we mainly compare our results |
|
with two fine-tune-based knowledge-enhanced models: KT-NET and SKG, which also evaluate |
|
their results on ReCoRD with BERT largeas the encoder part. As mentioned in the original paper |
|
of KT-NET, KT-NET mainly focuses on the extractive-style MRC task. We also evaluate KT-NET |
|
on the multiple-response-items-style MRC task and compare the results with KELM. We evaluate |
|
our approach in three different KB settings: KELM WordNet ,KELM NELL, and KELM Both, to inject KG |
|
from WordNet, NELL, and both of the two, respectively (The same as KT-NET). Implementation |
|
details of our model are presented in Appendix C. |
|
Dev Test |
|
Model EM F1 EM F1 |
|
BERT large 70.2 72.2 71.3 72.0 |
|
SKG+BERT large 70.9 71.6 72.2 72.8 |
|
KT-NET WordNet 70.6 72.8 - - |
|
KT-NET NELL 70.5 72.5 - - |
|
KT-NET BOTH 71.6 73.6 73.0 74.8 |
|
KELM WordNet 75.4 75.9 75.9 76.5 |
|
KELM NELL 74.8 75.3 75.9 76.3 |
|
KELM Both 75.1 75.6 76.2 76.7 |
|
Table 1: Result on ReCoRD.Dev Test |
|
Model EM F1 EM F1 |
|
BERT large - - 24.1 70.0 |
|
KT-NET |
|
BOTH 26.7 71.7 25.4 71.1 |
|
KELM WordNet 29.2 70.6 25.9 69.2 |
|
KELM NELL 27.3 70.4 26.5 70.6 |
|
KELM Both 30.3 71.0 27.2 70.8 |
|
Table 2: Result on MultiRC. [*] are from our |
|
implementation. |
|
4.3 R ESULTS |
|
The results for the extractive-style MRC task and multiple-response-items-style MRC task are given in |
|
Table 1 and Table 2, respectively. The scores of other models are taken directly from the leaderboard |
|
of SuperGLUE6and literature (Qiu et al., 2019; Yang et al., 2019). In this paper, our implementation |
|
is based on a single model, and hence comparing with ensemble based models is not considered. Best |
|
results are labeled in bold and the second best are underlined. |
|
Results on the dev set of ReCoRD show that: (1) KELM outperforms BERT large, irrespective of |
|
which external KG is used. Our best KELM offers a 5.2/3.7 improvement in EM/F1 over BERT large. |
|
(2) KELM outperforms previous SOTA knowledge-enhanced PLM (KT-NET) by +3.8 EM/+2.3 F1 . |
|
In addition, KELM outperforms KT-NET significantly in all three KB settings. On the dev set of |
|
MultiRC, the best KELM offers a 3.6improvement in EM over KT-NET. Although the performance |
|
on F1 drop a little compared with KT-NET, we still get a gain of +2.9 (EM+F1) over the former |
|
SOTA model7. |
|
Results on the test set further demonstrate the effectiveness of KELM and its superiority over the |
|
previous works. On ReCoRD, it significantly outperforms the former SOTA knowledge-enhanced |
|
PLM (finetuning based model) by +3.2 EM/+1.9 F1 . And on MultiRC, KELM offers a 3.1/0.8 |
|
improvement in EM/F1 over BERT large, and achieves a gain of +1.5 (EM+F1) over KT-NET. |
|
6https://super.gluebenchmark.com/leaderboard (Nov.14th, 2021) |
|
7The best model is chosen according to the EM+F1 score (same as KT-NET). |
|
7Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
5 C ASE STUDY |
|
This section uses an example in ReCoRD to show how KELM avoids knowledge ambiguity issue |
|
and selects the most relevant mentioned entities adaptively w.r.t the textual context. Recall that given |
|
a tokenti, the importance of a mentioned entity mq |
|
jinq-th KG is scored by the attention weight |
|
q |
|
ijin Eq.2. To illustrate how KELM can select the most relevant mentioned entities, we analyze |
|
the example that was also used in the case study part of KT-NET. The question of this example is |
|
“Sudan remains a XXX-designated state sponsor of terror and is one of six countries subject to the |
|
Trump administration’s ban”, where the “XXX” is the answer that needs to be predicted. The case |
|
study in KT-NET shows the top 3 most relevant concepts from WordNet for the word “ban” are |
|
“forbidding.n.01”, “proscription.n.01”, and “ban.v.02”, with the weights of 0.861, 0.135, and 0.002, |
|
respectively. KT-NET treats all synsets of a word as candidate KG concepts, both “forbidding.n.01” |
|
and “ban.v.02” will be the related concepts of the word “ban” in the text. Although KT-NET can |
|
select relevant concepts and suppress the knowledge noise through its specially designed attention |
|
mechanism, we still observe two problems from the previous case study: (1) KT-NET cannot select |
|
the most relevant mentioned entities in KG that share the same string in the input text. (2) Lack |
|
of ability to judge the part of speech (POS) of the word (e.g. “ban.v.02” gets larger weights than |
|
“ban.n.04”). |
|
Word in text |
|
(prototype)The most relevant |
|
mentioned entity in |
|
WordNet (predicted)Golden mentioned entity |
|
ford ford.n.05 (0.56) ford.n.05 |
|
pardon pardon.v.02 (0.86) pardon.v.02 |
|
nixon nixon.n.01 (0.74) nixon.n.01 |
|
lead lead.v.03 (0.73) lead.v.03 |
|
outrage outrage.n.02 (0.62) outrage.n.02 |
|
Table 3: Case study. Comparisons between the golden label |
|
with the most relevant mentioned entity in WordNet. The |
|
importance of selected mentioned entities is provided in the |
|
parenthesis.For KELM, by contrast, we focus on |
|
selecting the most relevant mentioned |
|
entities to solve the knowledge ambi- |
|
guity issue (based on the “entity link |
|
to token graph” part of UKET). For in- |
|
jecting WordNet, by allowing message |
|
passing on the extracted sub-graphs |
|
(“KG only” part of UKET), knowl- |
|
edge context can be dynamically em- |
|
bedded according to the textual con- |
|
text. Thus the neighbors’ information |
|
of mentioned entities in WordNet can |
|
be used to help the word in a text to |
|
correspond to a particular POS based on its context. The top 3 most relevant mentioned entities in |
|
WordNet for the word “ban” in the above example are “ban.n.04”, “ban.v.02”, and “ban.v.01”, with |
|
the weights of 0.715, 0.205, and 0.060, respectively. |
|
To vividly show the effectiveness of KELM, we analyze ambiguous words in the motivating example |
|
show in Fig. 1 (The example comes from ReCoRD): |
|
“President Ford then pardoned Richard Nixon, leading to a further firestorm of outrage. ” |
|
Table. 3 presents 5 words in the above passage. For each word, the most relevant mentioned entity in |
|
WordNet with the highest score is given. The golden mentioned entity for each word is labeled by |
|
us. Definitions of mentioned entities in WordNet that correspond to the word examples are listing in |
|
Table 5 of Appendix. |
|
6 C ONCLUSION |
|
In this paper, we have proposed KELM for MRC, which enhances PLM representations with |
|
structured knowledge from KGs based on the fine-tuning process. Via a unified knowledge-enhanced |
|
text graph, KELM can embed the injected knowledge dynamically, and select relevant mentioned |
|
entities in the input KGs. In the empirical analysis, KELM shows the effectiveness of fusing external |
|
knowledge into representations of PLM and demonstrates the ability to avoid knowledge ambiguity |
|
issue. Injecting emerging factual knowledge into PLM during finetuning without re-pretraining |
|
the whole model is quite important in the application of PLMs and is still barely investigated. |
|
Improvements achieved by KELM over vanilla baselines indicate a potential direction for future |
|
research. |
|
8Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
ACKNOWLEDGEMENTS |
|
The authors thank Ms. X. Lin for insightful comments on the manuscript. We also thank Dr. Y . Guo |
|
for helpful suggestions in parallel training settings. We also thank all the colleagues in AI Application |
|
Research Center (AARC) of Huawei Technologies for their supports. |
|
REFERENCES |
|
Antoine Bordes, Nicolas Usunier, Alberto García-Durán, J. Weston, and Oksana Yakhnenko. Trans- |
|
lating embeddings for modeling multi-relational data. In NIPS , 2013. |
|
A. Carlson, J. Betteridge, B. Kisiel, B. Settles, E.R. Hruschka Jr., and T.M. Mitchell. Toward an |
|
architecture for never-ending language learning. In Proceedings of the Conference on Artificial |
|
Intelligence (AAAI) , pp. 1306–1313. AAAI Press, 2010. |
|
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. BERT: Pre-training of deep |
|
bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of |
|
the North American Chapter of the Association for Computational Linguistics: Human Language |
|
Technologies, Volume 1 (Long and Short Papers) , pp. 4171–4186, Minneapolis, Minnesota, June |
|
2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1423. URL https: |
|
//www.aclweb.org/anthology/N19-1423 . |
|
P. Ferragina and Ugo Scaiella. Tagme: on-the-fly annotation of short text fragments (by wikipedia |
|
entities). Proceedings of the 19th ACM international conference on Information and knowledge |
|
management , 2010. |
|
Lu Haonan, Seth H. Huang, Tian Ye, and Guo Xiuyan. Graph star net for generalized multi-task |
|
learning, 2019. |
|
Daniel Khashabi, Snigdha Chaturvedi, Michael Roth, Shyam Upadhyay, and Dan Roth. Looking |
|
beyond the surface:a challenge set for reading comprehension over multiple sentences. In Pro- |
|
ceedings of North American Chapter of the Association for Computational Linguistics (NAACL) , |
|
2018. |
|
Thomas N. Kipf and Max Welling. Semi-supervised classification with graph convolutional networks, |
|
2017. |
|
Weijie Liu, Peng Zhou, Zhe Zhao, Zhiruo Wang, Qi Ju, Haotang Deng, and Ping Wang. K-bert: |
|
Enabling language representation with knowledge graph, 2019a. |
|
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike |
|
Lewis, Luke Zettlemoyer, and Veselin Stoyanov. Roberta: A robustly optimized bert pretraining |
|
approach, 2019b. |
|
Edward Loper and Steven Bird. Nltk: The natural language toolkit. In Proceedings of the ACL-02 |
|
Workshop on Effective Tools and Methodologies for Teaching Natural Language Processing and |
|
Computational Linguistics - Volume 1 , ETMTNLP ’02, pp. 63–70, USA, 2002. Association for |
|
Computational Linguistics. doi: 10.3115/1118108.1118117. URL https://doi.org/10. |
|
3115/1118108.1118117 . |
|
Haonan Lu and Hailin Hu. Dense: An enhanced non-abelian group representation for knowledge |
|
graph embedding, 2020. |
|
George A. Miller. Wordnet: A lexical database for english. COMMUNICATIONS OF THE ACM , 38: |
|
39–41, 1995. |
|
Matthew E. Peters, Mark Neumann, Robert L. Logan IV au2, Roy Schwartz, Vidur Joshi, Sameer |
|
Singh, and Noah A. Smith. Knowledge enhanced contextual word representations, 2019. |
|
Nina Poerner, Ulli Waltinger, and Hinrich Schütze. Bert is not a knowledge base (yet): Factual |
|
knowledge vs. name-based reasoning in unsupervised qa. ArXiv , abs/1911.03681, 2019. |
|
9Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
Delai Qiu, Yuanzhe Zhang, Xinwei Feng, Xiangwen Liao, Wenbin Jiang, Yajuan Lyu, Kang Liu, and |
|
Jun Zhao. Machine reading comprehension using structural knowledge graph-aware network. In |
|
Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and |
|
the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) , pp. |
|
5896–5901, Hong Kong, China, November 2019. Association for Computational Linguistics. doi: |
|
10.18653/v1/D19-1602. URL https://www.aclweb.org/anthology/D19-1602 . |
|
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi |
|
Zhou, Wei Li, and Peter J. Liu. Exploring the limits of transfer learning with a unified text-to-text |
|
transformer, 2020. |
|
Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang. Squad: 100,000+ questions for |
|
machine comprehension of text, 2016. |
|
Melissa Roemmele, Cosmin Adrian Bejan, and Andrew S. Gordon. Choice of plausible alternatives: |
|
An evaluation of commonsense causal reasoning. In Logical Formalizations of Commonsense |
|
Reasoning, Papers from the 2011 AAAI Spring Symposium, Technical Report SS-11-06, Stanford, |
|
California, USA, March 21-23, 2011 . AAAI, 2011. URL http://www.aaai.org/ocs/ |
|
index.php/SSS/SSS11/paper/view/2418 . |
|
Michael Schlichtkrull, Thomas N. Kipf, Peter Bloem, Rianne van den Berg, Ivan Titov, and Max |
|
Welling. Modeling relational data with graph convolutional networks, 2017. |
|
Yusheng Su, Xu Han, Zhengyan Zhang, Peng Li, Zhiyuan Liu, Yankai Lin, Jie Zhou, and Maosong |
|
Sun. Cokebert: Contextual knowledge selection and embedding towards enhanced pre-trained |
|
language models, 2020. |
|
Tianxiang Sun, Yunfan Shao, Xipeng Qiu, Qipeng Guo, Yaru Hu, Xuanjing Huang, and Zheng Zhang. |
|
Colake: Contextualized language and knowledge embedding, 2020. |
|
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz |
|
Kaiser, and Illia Polosukhin. Attention is all you need, 2017. |
|
Petar Veli ˇckovi ´c, Guillem Cucurull, Arantxa Casanova, Adriana Romero, Pietro Liò, and Yoshua |
|
Bengio. Graph attention networks, 2018. |
|
Alex Wang, Yada Pruksachatkun, Nikita Nangia, Amanpreet Singh, Julian Michael, Felix Hill, Omer |
|
Levy, and Samuel R. Bowman. Superglue: A stickier benchmark for general-purpose language |
|
understanding systems, 2020a. |
|
Minjie Wang, Da Zheng, Zihao Ye, Quan Gan, Mufei Li, Xiang Song, Jinjing Zhou, Chao Ma, |
|
Lingfan Yu, Yu Gai, Tianjun Xiao, Tong He, George Karypis, Jinyang Li, and Zheng Zhang. Deep |
|
graph library: A graph-centric, highly-performant package for graph neural networks, 2020b. |
|
Ruize Wang, Duyu Tang, Nan Duan, Zhongyu Wei, Xuanjing Huang, Jianshu ji, Guihong Cao, Daxin |
|
Jiang, and Ming Zhou. K-adapter: Infusing knowledge into pre-trained models with adapters, |
|
2020c. |
|
Xiaozhi Wang, Tianyu Gao, Zhaocheng Zhu, Zhengyan Zhang, Zhiyuan Liu, Juanzi Li, and Jian |
|
Tang. Kepler: A unified model for knowledge embedding and pre-trained language representation, |
|
2020d. |
|
Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Clement Delangue, Anthony Moi, |
|
Pierric Cistac, Tim Rault, Rémi Louf, Morgan Funtowicz, Joe Davison, Sam Shleifer, Patrick von |
|
Platen, Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu, Teven Le Scao, Sylvain Gugger, Mariama |
|
Drame, Quentin Lhoest, and Alexander M. Rush. Huggingface’s transformers: State-of-the-art |
|
natural language processing, 2020. |
|
Wenhan Xiong, Jingfei Du, William Yang Wang, and Veselin Stoyanov. Pretrained encyclopedia: |
|
Weakly supervised knowledge-pretrained language model, 2019. |
|
Ikuya Yamada, Akari Asai, Hiroyuki Shindo, Hideaki Takeda, and Yuji Matsumoto. Luke: Deep |
|
contextualized entity representations with entity-aware self-attention, 2020. |
|
10Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
An Yang, Quan Wang, Jing Liu, Kai Liu, Yajuan Lyu, Hua Wu, Qiaoqiao She, and Sujian Li. |
|
Enhancing pre-trained language representations with rich knowledge for machine reading com- |
|
prehension. In Proceedings of the 57th Annual Meeting of the Association for Computational |
|
Linguistics , pp. 2346–2357, Florence, Italy, July 2019. Association for Computational Linguistics. |
|
doi: 10.18653/v1/P19-1226. URL https://www.aclweb.org/anthology/P19-1226 . |
|
Bishan Yang and Tom Mitchell. Leveraging knowledge bases in LSTMs for improving machine |
|
reading. In Proceedings of the 55th Annual Meeting of the Association for Computational Lin- |
|
guistics (Volume 1: Long Papers) , pp. 1436–1446, Vancouver, Canada, July 2017. Association for |
|
Computational Linguistics. doi: 10.18653/v1/P17-1132. URL https://www.aclweb.org/ |
|
anthology/P17-1132 . |
|
Bishan Yang, Wen tau Yih, Xiaodong He, Jianfeng Gao, and Li Deng. Embedding entities and |
|
relations for learning and inference in knowledge bases, 2015. |
|
Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Ruslan Salakhutdinov, and Quoc V . Le. |
|
Xlnet: Generalized autoregressive pretraining for language understanding, 2020. |
|
Donghan Yu, Chenguang Zhu, Yiming Yang, and Michael Zeng. Jaket: Joint pre-training of |
|
knowledge graph and language understanding, 2020. |
|
Sheng Zhang, Xiaodong Liu, Jingjing Liu, Jianfeng Gao, Kevin Duh, and Benjamin Van Durme. |
|
Record: Bridging the gap between human and machine commonsense reading comprehension, |
|
2018. |
|
Zhengyan Zhang, Xu Han, Zhiyuan Liu, Xin Jiang, Maosong Sun, and Qun Liu. Ernie: Enhanced |
|
language representation with informative entities, 2019. |
|
ASUMMARY AND COMPARISON OF RECENT KNOWLEDGE -ENHANCED PLM S |
|
Table 4 shows a brief summary and comparison of recent knowledge-enhanced PLMs. Most of recent |
|
work concentrated on injecting external knowledge graphs during pre-training phase, which makes |
|
them inefficient in injecting external knowledge (e.g. LUKE takes about 1000 V100 GPU days to |
|
re-pretraining the RoBERTa based PLM model). Also, nearly all of them uses an additional entity |
|
linking tool to align the mentioned entities in the Wikidata to the entity mentions in the pre-trained |
|
corpus (English Wikipedia) uniquely. These methods never consider to resolve the knowledge |
|
ambiguity problem. |
|
B D ATATSET AND KNOWLEDGE GRAPH DETAILS |
|
ReCoRD (an acronym for the Reading Comprehension with Commonsense Reasoning Dataset) is a |
|
large-scale dataset for extractive-style MRC requiring commonsense reasoning. There are 100,730, |
|
10,000, and 10,000 examples in the training, development (dev), and test set, respectively. An example |
|
of the ReCoRD consists of three parts: passage, question, and answer. The passage is formed by |
|
the first few paragraphs of an article from CNN or Daily Mail, with named entities recognized and |
|
marked. The question is a sentence from the rest of the article, with a missing entity specified as the |
|
golden answer. The model needs to find the golden answer among the entities marked in the passage. |
|
Questions that can be easily answered by pattern matching are filtered out. By the design of the |
|
process of data collection, one can see that to answer the questions, external background knowledge |
|
and ability of reasoning are required. |
|
MultiRC (Multi-Sentence Reading Comprehension) is a multiple-response-items-style MRC dataset |
|
of short paragraphs and multi-sentence questions that can be answered from the content of the |
|
paragraph. Each example of MultiRC includes a question that associates with several choices for |
|
answer-options, and the number of correct answer-options is not pre-specified. The correct answer is |
|
not required to be a span in the text. The dataset consists of 10K questions ( 6k multiple-sentence |
|
questions), about 60% of this data make training/dev data. Paragraphs in the dataset have diverse |
|
provenance by being extracted from 7 different domains such as news, fiction, historical text etc., and |
|
hence are expected to be more complicated in their contents as compared to single-domain datasets. |
|
11Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
Model Downstream Task Used KGsNeed |
|
Pre-trainDynamically |
|
Embedding KG |
|
ContextInject external |
|
KG’s RepresentationsSupport |
|
Multi-relationalSupport |
|
Multi-hopHandle Knowledge |
|
Ambiguity IssueBase Model |
|
ERNIE (Zhang et al., 2019)Glue, Ent Typing |
|
Rel CLSWikidataYes |
|
(MLM, NSP, |
|
Ent Mask task)NoInject pretrained |
|
entity embeddings |
|
(TransE) explicitlyNo |
|
(only entity embedding)NoNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)BERT base |
|
K-BERT (Liu et al., 2019a)Q&A, NER |
|
Sent CLSCN-DBpedia |
|
HowNet, MedicalKGOptional |
|
(MLM, NSP)No NoYes |
|
(treat relations as words)NoNo |
|
(designed ATT mechanism |
|
can solve KN issue)BERT base |
|
KnowBERT (Peters et al., 2019)Rel Extraction |
|
Ent TypingCrossWikis, |
|
WordNetYes |
|
(MLM, NSP, |
|
Ent Linking task)NoInject both pretrained |
|
entity embeddings (TuckER) |
|
and entity definition explicitlyNo |
|
(only entity embedding)NoYes |
|
(weighed entity embeddings |
|
shared the same text)BERT base |
|
WKLM (Xiong et al., 2019) Q&A, Ent Typing WikidataYes |
|
(MLM, Ent |
|
replacement task)No No No NoNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)BERT base |
|
K-Adapter (Wang et al., 2020c)Q&A, |
|
Ent TypingWikidata |
|
Dependency ParsingYes |
|
(MLM, |
|
Rel predition task)No NoYes |
|
(Via Rel prediction task |
|
during pretraining)NoNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)RoBERTa large |
|
KEPLER (Wang et al., 2020d)Ent Typing |
|
Glue, Rel CLS |
|
Link PredictionWikidataYes |
|
(MLM, |
|
Link predition task)YesInject embeddings of |
|
entity and relation’s |
|
description explicitlyYes |
|
(Via link prediction task |
|
during pretraining)NoNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)RoBERTa base |
|
JAKET (Yu et al., 2020)Rel CLS, KGQA |
|
Ent CLSWikidataYes |
|
(MLM, Ent Mask task, |
|
Ent category prediction, |
|
Rel type prediction)YesInject embeddings of |
|
entity descriptionsYes |
|
(Via Rel type prediction |
|
during pretraining)YesNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)RoBERTa base |
|
CoLAKE (Sun et al., 2020)Glue, Ent Typing |
|
Rel ExtractionWikidataYes |
|
(MLM, Ent Mask task, |
|
Rel type prediction)Yes NoYes |
|
(treat relations as words)NoNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)RoBERTa base |
|
LUKE (Yamada et al., 2020)Ent Typing, Rel CLS |
|
NER, Q&AEnt from |
|
WikipediaYes |
|
(MLM, Ent Mask task)No No No NoNo |
|
(treat hyperlinks in Wikipedia |
|
as entity annotations)RoBERTa large |
|
CokeBERT (Su et al., 2020)Rel CLS |
|
Ent TypingWikidataYes |
|
(MLM, NSP, |
|
Ent Mask task)YesInject pretrained |
|
entity embeddings |
|
(TransE) explicitlyYes |
|
(Via S-GNN to encode KG |
|
context dynamically)YesNo |
|
(anchored entity mention to |
|
the unique id of Wikidata)RoBERTa large |
|
SKG (Qiu et al., 2019) MRC WordNet, ConceptNet No YesInject pretrained |
|
entity embeddings |
|
(BILINER) explicitlyYes |
|
(Via multi-relational |
|
GNN to encode KG |
|
context dynamically)Yes No BERT large |
|
KT-NET (Yang et al., 2019) MRC WordNet, NELL No NoInject pretrained |
|
entity embeddings |
|
(BILINER) explicitlyNo |
|
(only entity embedding)NoYes |
|
(dynamically selecting |
|
KG context)BERT large |
|
KELM MRC WordNet, NELL No YesInject pretrained |
|
entity embeddings |
|
(BILINER) explicitlyYes |
|
(Via multi-relational |
|
GNN to encode KG |
|
context dynamically)YesYes |
|
(dynamically selecting |
|
related mentioned entity)BERT large |
|
Table 4: A brief summary and comparison of recent knowledge-enhanced PLMs. The full names |
|
of some abbreviations are as follows. MLM : masked language model, NSP: next sentence pre- |
|
diction, Ent: entity, Rel: relation, CLS : classification, Sent: sentence, ATT : attention. Com- |
|
ments/descriptions of features are written in parentheses. Desired properties are written in bold . |
|
WordNet contains 151,442 triplets with 40,943 synsets459and 18 relations. We look up mentioned |
|
entities in the WordNet by string matching operation, and link all tokens in the same word to the |
|
retrieved mentioned entities (tokens are tokenized by Tokenizer of BERT). Then, we extract all 1-hop |
|
neighbors for each mentioned entity and construct sub-graphs. In this paper, our experiment results |
|
are based on the 1-hop case. However, our framework can be generalized to multi-hop easily, and we |
|
leave this for future work. |
|
NELL contains 180,107 entities and 258 concepts. We link entity mentions to the whole KG, and |
|
return associated concepts. |
|
C I MPLEMENTATION DETAILS |
|
Our implementation is based on HuggingFace (Wolf et al., 2020) and DGL (Wang et al., 2020b). |
|
For all three settings of KELM, parameters of the encoding layer of BERT largeare initialized with |
|
pre-trained model released by Google. Other trainable parameters in HMP are randomly initialized. |
|
The total number of trainable parameters of KELM is 340.4M (Roughly the same as BERT large, which |
|
has 340M parameters). Since including all neighbors around mentioned entities of WordNet is not |
|
efficient, for simplicity, we use top 3 most common relations in WordNet in our experiment (i.e. |
|
hyponym, hypernym, derivationally_related_form). For both datasets, we use a “two stage” fine-tune |
|
strategy to achieve our best performance, the FullTokenizer built in BERT is used to segment input |
|
words into wordpieces. |
|
For ReCoRD, the maximum length of answer during inference is set to 30, and the maximum length |
|
of question is set to 64. Questions longer than that are truncated. The maximum length of input |
|
sequenceT8is set to 384. Input sequences longer than that are segmented into chunks with a stride |
|
of 128. Fine-tuning our model on ReCoRD costs about 18 hours on 4 V100 GPU with a batch size of |
|
48. We freeze parameters of BERT and use Adam optimizer with a learning rate of 1e-3 to train our |
|
knowledge module in the first stage. The maximum number of training epochs of the first stage is |
|
10. The purpose of this is to provide a good weight initialization for our HMP. For the second stage, |
|
the pre-trained BERT parameters and our HMP part will be fine-tuned together. The max number of |
|
training epochs is chosen from f4;6;8g. The learning rate is set to be 2e-5 with a warmup over the |
|
8Refer to the PLM Encoding Module of Methodology Section. |
|
12Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
first6%of max steps, and linear decay until up to max epochs. For both two stages, early stopping is |
|
applied according to the best EM+F1 score on the dev set every 500 steps. |
|
For MultiRC, the maximum length of input sequence Tis set to 256. The summation of length of |
|
question (Q) and length of candidate answer ( A) is not limited. Paragraph ( P) is truncated to fit the |
|
maximum length of input sequence. Fine-tuning KELM on MultiRC needs about 12 hours on 4 V100 |
|
GPU with a batch size of 48. For the first stage finetuning, learning rate is 1e-4 and the maximum |
|
number of training epochs is 10. For the second stage, the max number of training steps is chosen |
|
fromf10000;15000;20000g. The learning rate is set to be 2e-5 with a warmup over the first 10% of |
|
max steps. |
|
D S UPPLEMENTATION OF THE CASE STUDY SECTION |
|
We provide definitions of the top 3 most relevant mentioned entities in WordNet that correspond to the |
|
word examples mentioned in Case Study Section . Descriptions are obtained by using NLTK (Loper |
|
& Bird, 2002). By comprehending the motivating example in the case study section, we can see that |
|
KELM can correctly select the most relevant mentioned entities in the KG. |
|
Word in text |
|
(prototype)Mentioned entity |
|
in WordNetDefinition |
|
banban.n.04 (0.72) an official prohibition or edict against something |
|
ban.v.02 (0.21) prohibit especially by legal means or social pressure |
|
ban.v.01 (0.06) forbid the public distribution of ( a movie or a newspaper) |
|
fordford.n.05 (0.56)38th President of the United States; |
|
appointed vice president and succeeded |
|
Nixon when Nixon resigned (1913-) |
|
ford.n.07 (0.24) a shallow area in a stream that can be forded |
|
ford.v.01 (0.08) cross a river where it’s shallow |
|
pardonpardon.v.02 (0.86) a warrant granting release from punishment for an offense |
|
sentinel (0.10) - |
|
pardon.n.02 (0.04) grant a pardon to |
|
nixonnixon.n.01 (0.74)vice president under Eisenhower and 37th President |
|
of the United States; resigned after the Watergate |
|
scandal in 1974 (1913-1994) |
|
sentinel (0.26) - |
|
leadlead.v.03 (0.73) tend to or result in |
|
lead.n.03 (0.12) evidence pointing to a possible solution |
|
lead.v.04 (0.05) travel in front of; go in advance of others |
|
outrageoutrage.n.02 (0.62) a wantonly cruel act |
|
sentinel (0.38) - |
|
Table 5: Definitions of mentioned entities in WordNet corresponding to the word examples in the case |
|
study. The importance of mentioned entities is provided in the parenthesis. “sentinel” is meaningless, |
|
which is used to avoid knowledge noise. |
|
E E XPERIMENT ON COMMONSENSE CAUSAL REASONING TASK |
|
To further explore the generalization ability of KELM, we also evaluate our method on COPA (Roem- |
|
mele et al., 2011) (Choice of Plausible Alternatives), which is also a benchmark dataset in SuperGLUE |
|
and can be used for evaluating progress in open-domain commonsense causal reasoning. COPA |
|
consists of 1000 questions, split equally into development and test sets of 500 questions each. Each |
|
question is composed of a premise and two alternatives, where the task is to select the alternative |
|
that more plausibly has a causal relation with the premise. Similar to the previous two MRC tasks, |
|
the development set is publicly available, but the test set is hidden. One has to submit the predicted |
|
results for the test set to SuperGLUE to retrieve the final test score. Since the implementation of |
|
KELM is based on BERT large, we use it as our baseline for the comparison. The result of BERT large |
|
is directly taken from the leaderboard of SuperGLUE. Table 6 shows the experiment results. The |
|
injected KG is WordNet here, and we use accuracy as the evaluation metric. |
|
The huge improvement over the baseline in this task demonstrates that knowledge in WordNet is |
|
indeed helpful for BERT to improve the generalization ability to the out-of-domain downstream task. |
|
13Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
Model dev test |
|
BERT large - 70.6 |
|
KELMBERTlarge |
|
WordNet76.1 78.0 |
|
Table 6: Performance comparison on COPA. The effectiveness of injecting knowledge (WordNet) are |
|
shown. |
|
F KELM: A FRAMEWORK OF FINETUNE -BASED MODEL -AGNOSTIC |
|
KNOWLEDGE -ENHANCED PLM |
|
We implement KELM based on the RoBERTa large, which has a similar number of trainable parameters |
|
asBERT largebut uses nearly 10 times of training corpus than BERT large. Since the performances of |
|
RoBEATa on the leaderboard of SuperGLUE are based on ensembling, we also finetune RoBERTa large |
|
on ReCoRD to produce the results of a single model. Comparisons of the results can be found |
|
in Table 7, where you can also see an improvement there. However, that improvement is not |
|
as significant as we observed in BERT large. Reasons are two-fold: (1) Passages in ReCoRD are |
|
collected from articles in CNN/Daily Mail, while BERT is pre-trained on BookCorpus and English |
|
Wikipedia. RoBERTa not only uses the corpus that used in BERT (16 GB), but also an additional |
|
corpus collected from the CommonCrawl News dataset (76 GB). ReCoRD dataset is in-domain |
|
for RoBERTa but is out-of-domain for BERT. It seems that the improvements of KELM with |
|
injecting general KGs (e.g. WordNet) on the in-domain downstream tasks are not as large |
|
as the out-of-domain downstream tasks . A similar phenomenon can be also observed in the |
|
experiment of SQuAD 1.1 (Refer to Appendix E). (2) The same external knowledge (WordNet, |
|
NELL) can not help RoBERTa largetoo much, since RoBERTa is pre-trained on a much larger corpus |
|
than BERT, knowledge in WordNet/NELL has been learned in RoBERTa. |
|
Dev Test |
|
Model EM F1 EM F1 |
|
PLM w/o |
|
external knowledgeBERT large 70.2 72.2 71.3 72.0 |
|
RoBERTa |
|
large 87.9 88.4 88.4 88.9 |
|
knowledge enhanced PLM |
|
(finetune-based)KELMBERTlarge |
|
Both75.1 75.6 76.2 76.7 |
|
KELMRoBERTalarge |
|
Both88.2 88.7 89.1 89.6 |
|
knowledge enhanced PLM |
|
(pretrain-based)LUKE 90.8 91.4 90.6 91.2 |
|
Table 7: Comparison of the effectiveness of injecting external knowledge between BERT and |
|
RoBERTa. [*] Results are from our implementation. |
|
We also list the results of LUKE (Yamada et al., 2020) in Table 7. LUKE is a pretrain-based |
|
knowledge enhanced PLM and uses Wiki-related golden entities (one-to-one mapping) as the injected |
|
knowledge source (about 500k entities9). It has more 128 M parameters than the total number |
|
of parameters of the vanilla RoBERTa. As we summarized in Table 4 in the main text, the pre- |
|
training task is also different compared with RoBERTa. Although LUKE gets better results compared |
|
with vanilla RoBERTa and KELM, it needs 16 NVIDIA Tesla V100 GPUs and the training takes |
|
approximately 30 days. Relying on hyperlinks in Wikipedia as golden entity annotations, lacking |
|
the flexibility to adapt the external knowledge of other domains, and needing re-pretraining when |
|
incorporating knowledge, these limitations hinder the abilities of applications. |
|
G E XPERIMENT ON SQUAD 1.1 |
|
SQuAD1.1 (Rajpurkar et al., 2016) is a well known extractive-style MRC dataset that consists of |
|
questions created by crowdworkers for Wikipedia articles . It contains 100,000+ question-answer |
|
pairs on 536 articles. We implement KELM based on the BERT large, and compare our results on the |
|
development set of SQuAD 1.1 with KT-NET (Best result of KT-NET is based on injecting WordNet |
|
only). Results are shown in Table 8 |
|
9For KELM, we only use 40943 entities in WordNet and 258 concepts in NELL. |
|
14Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
Dev |
|
Model EM F1 |
|
PLM w/o external knowledge BERT large 84.4 91.2 |
|
knowledge enhanced PLM |
|
(finetune-based)KT-NETBERTlarge |
|
WordNet85.1 91.7 |
|
KELMBERTlarge |
|
WordNet84.7 91.5 |
|
Table 8: Performance comparison on the development set of SQuAD 1.1. |
|
Results on KELM show an improvement over vanilla BERT. Both BERT and RoBERTa use English |
|
Wikipedia as the corpus for pretraining. Since SQuAD is also created from Wikipedia, it is an |
|
in-domain downstream task for both BERT and RoBERTa (while ReCoRD dataset is in-domain for |
|
RoBERTa but is out-of-domain for BERT). This explains why RoBERTa achieves a much larger |
|
improvement over BERT on the result of ReCoRD ( 71:3!88:4in EM on test set) than the one |
|
on SQuAD 1.1 ( 84:1!88:9). The rest of the improvement is because RoBERTa uses 10 times of |
|
training corpus than BERT and different pre-training strategies they used. |
|
Interestingly, we find the performance of KELM on SQuAD 1.1 is sub-optimal compared with KT- |
|
NET. As we mentioned in the last paragraph of the Related Work Section , KT-NET treats all synsets |
|
of entity mentions within the WN18 as candidate KB concepts. Via a specially designed attention |
|
mechanism, KT-NET can directly use all 1-hop neighbors of the mentioned entities. Although this |
|
limits the ability of KT-NET to select the most relevant mentioned entities (as we discussed in Case |
|
Study Section ), information of these neighbors can be directly considered. Using neighbors of the |
|
mentioned entities indirectly via the HMP mechanism makes it possible for KELM to dynamically |
|
embed injected knowledge and to select semantics-related mentioned entities. However, SQuAD |
|
is an in-domain downstream task for BERT, the problem of ambiguous meanings of words can be |
|
alleviated by pretraining model on the in-domain corpus. Compared with KT-NET, a longer message |
|
passing path in KELM may lead to sub-optimal improvement on the in-domain task. |
|
H F URTHER DISCUSSIONS ABOUT THE NOVELTY W .R.TSKG/KT-NET |
|
UKET defined in KELM consists of three subgraphs in a hierarchical structure, each subgraph |
|
corresponds to one sub-process of our proposed HMP mechanism and solves one problem presented |
|
in the Hierarchical Knowledge Enhancement Module part of Methodology Section . SKG only |
|
uses GNN to dynamically encode the extracted KG which corresponds to the first part of UKET, it can |
|
not solve the knowledge ambiguity issue and forbids interactions among knowledge-enriched tokens. |
|
KT-NET defines a similar graph as the third part of UKET. However, the first and second subgraphs |
|
of UKET are absent. The second subgraph of UKET is independent of ideas of KT-NET and SKG, |
|
thus KELM is not a simple combination of these two methods. We are the first to unify text and |
|
KG into a graph and to propose this hierarchical message passing framework to incorporate two |
|
heterogeneous information. SKG/KT-NET can be interpreted as parts of the ablation study of |
|
components of KELM . The result of SKG is ablation with the component only related to the first |
|
subgraph of UKET. While KT-NET only contains the third subgraph with a modified knowledge |
|
integration module. KELM uses a dedicatedly designed HMP mechanism to let the information of |
|
farther neighbors to be considered. However, longer information passing path than KT-NET makes |
|
it less efficient. In our experiments, KELM takes more 30% training time than KT-NET on both |
|
ReCoRD and MultiRC. |
|
I L IMITATIONS AND FURTHER IMPROVEMENTS OF KELM |
|
Limitations for KELM are two-fold: (1) Meanings of mentioned entities in different KGs that share |
|
the same entity mentions in the text may conflict with each other. Although HMP can help to select |
|
the most relevant mentioned entities in a single KG, there’s no mechanism to guarantee the selections |
|
across different KGs; (2) Note the knowledge-enriched representation in Eq.3 is obtained by simple |
|
concatenation of the embeddings from different KGs. Too much knowledge incorporation may |
|
divert the sentence from its correct meaning (Knowledge noise issue). We expect these two potential |
|
improvements to be a promising avenue for future research. |
|
15Accepted at the ICLR 2022 Workshop on Deep Learning on Graphs for Natural Language Processing |
|
J F URTHER ANALYSIS AND DISCUSSION |
|
KELM incorporates knowledge in KGs into the representations in the last hidden layer of PLM |
|
(Refer to Methodology Section ). It is essentially a model-agnostic, KG-agnostic, and task-agnostic |
|
framework for enhancing language model representations with factual knowledge from KGs. It |
|
can be used to enhance any PLM, with any injected KGs, on any downstream task. Besides the |
|
two Q&A-related MRC tasks we mentioned in the main text, we also evaluate KELM on COPA |
|
and SQuAD 1.1 based on BERT large, results are presented in Appendix E and Appendix G, respec- |
|
tively. To demonstrate KELM is a model-agnostic framework, we also implement KELM based on |
|
RoBERTa largeand evaluate it on ReCoRD. The experiment is presented in Appendix F. Improvements |
|
achieved by KELM over all vanilla base PLM models indicate the effectiveness of injecting external |
|
knowledge. |
|
However, the improvements of KELM over RoBERTa on ReCoRD and BERT on SQuAD 1.1 are |
|
marginal compared with the ones on ReCoRD/MultiRC/COPA ( BERT largebased). The reason behind |
|
this is that pretraining model on in-domain unlabeled data could boost performance on downstream |
|
tasks. Passages in ReCoRD are collected from articles in CNN/Daily Mail, while BERT is pre-trained |
|
on BookCorpus and English Wikipedia. RoBERTa not only uses the corpus that used in BERT |
|
(16 GB), but also an additional corpus collected from the CommonCrawl News dataset (76 GB). |
|
ReCoRD is in-domain for RoBERTa but is out-of-domain for BERT. Similarly, SQuAD 1.1 is |
|
created from Wikipedia, it is an in-domain downstream task for both BERT and RoBERTa. This |
|
partially explains why RoBERTa achieves a much larger improvement over BERT on the result of |
|
ReCoRD ( 71:3!88:4in EM on test set) than the one on SQuAD 1.1 ( 84:1!88:9). A similar |
|
analysis can be also found in T5 (Raffel et al., 2020). From our empirical results, we can summarize |
|
that general KG (e.g. WordNet) can not help too much for the PLMs pretrained on in-domain data. |
|
But it can still improve the performance of the model when the downstream tasks are out-of-domain. |
|
Further detailed analysis can be found in our appendix. |
|
Finding a popular NLP task/dataset that is not related to the training corpus of modern PLMs is |
|
difficult. Pre-training on large-scale corpus is always good if we have unlimited computational |
|
resources and plenty of in-domain corpus. It has been evident that the simple finetuning of PLM is |
|
not sufficient for domain-specific applications. KELM can provide people another choice when they |
|
do not have such a large-scale in-domain corpus and want to incorporate incremental domain-related |
|
structural knowledge into the domain-specific applications. |
|
16 |