arxiv_dump / txt /2106.14625.txt
billxbf's picture
Upload 101 files
8f1929a verified
raw
history blame
34.7 kB
AMU-EURANOV A at CASE 2021 Task 1: Assessing the stability of
multilingual BERT
L´eo Bouscarrat1;2, Antoine Bonnefoy1, C´ecile Capponi2, Carlos Ramisch2
1EURA NOV A, Marseille, France
2Aix Marseille Univ, Universit ´e de Toulon, CNRS, LIS, Marseille, France
fleo.bouscarrat, antoine.bonnefoy [email protected]
fleo.bouscarrat, cecile.capponi, carlos.ramisch [email protected]
Abstract
This paper explains our participation in task 1
of the CASE 2021 shared task. This task is
about multilingual event extraction from news.
We focused on sub-task 4, event information
extraction. This sub-task has a small training
dataset and we fine-tuned a multilingual BERT
to solve this sub-task. We studied the instabil-
ity problem on the dataset and tried to mitigate
it.
1 Introduction
Event extraction is becoming more and more impor-
tant as the number of online news increases. This
task consists of extracting events from documents,
especially news. An event is defined by a group of
entities that give some information about the event.
Therefore, the goal of this task is to extract, for
each event, a group of entities that define the event,
such as the place and time of the event.
This task is related but still different from named
entity recognition (NER) as the issue is to group
the entities that are related to the same event, and
differentiate those related to different events. This
difference makes the task harder and also compli-
cates the annotation.
In the case of this shared task, the type of events
to extract is protests (H ¨urriyeto ˘glu et al., 2021a,b).
This shared task is in the continuation of two previ-
ous shared tasks at CLEF 2019 (H ¨urriyeto ˘glu et al.,
2019) and AESPEN (H ¨urriyeto ˘glu et al., 2020).
The first one deals with English event extraction
with three sub-tasks: document classification, sen-
tence classification, and event information extrac-
tion. The second focuses on event sentence co-
reference identification, whose goal is to group
sentences related to the same events.
This year, task 1 is composed of the four afore-
mentioned tasks and adds another difficulty: multi-
linguality. This year’s data is available in English,Spanish, and Portuguese. Thus, it is important to
note that there is much more data in English than
in the other languages. For the document classi-
fication sub-task, to test multilingual capabilities,
Hindi is available on the testing set only.
We have mainly focused on the last sub-task
(event information extraction), but we have also
submitted results for the first and second sub-tasks
(document and sentence classification). We used
multilingual BERT (Devlin et al., 2019), hence-
forth M-BERT, which is a model known to obtain
near state-of-the-art results on many tasks. It is also
supposed to work well for zero-or-few-shot learn-
ing on different languages (Pires et al., 2019). We
will see the results on these sub-tasks, especially
for sub-task 4 where the training set available for
Spanish and Portuguese is small.
Thus, one of the issues with transformer-based
models such as M-BERT is the instability on small
datasets (Dodge et al., 2020; Ruder, 2021). The
instability issue is the fact that by changing some
random seeds before the learning phase but using
the same architecture, data and hyper-parameters
the results can have a great variance. We will look
at some solutions to mitigate this issue, and how
this issue is impacting our results for sub-task 4.1
2 Tasks and data
Sub-tasks 1 and 2 can be seen as binary sequence
classification, where the goal is to say if a given
sequence is part of a specific class. In our case, a
classifier must predict whether a document contains
information about an event for sub-task 1 or if a
sentence contains information about an event for
sub-task 2.
Document and sentence classification tasks, sub-
tasks 1 and 2, are not our main research interest.
1Our code is available here: https://github.com/
euranova/AMU-EURANOVA-CASE-2021arXiv:2106.14625v2 [cs.CL] 4 Aug 2021Figure 1: Example of a snippet from sub-task 4.
Moreover, the datasets provided for these tasks
are less interesting (reasonable amount of training
data).
On the other hand, sub-task 4 not only has less
training data available but also requires more fine-
grained token-based prediction. The goal of sub-
task 4 is to extract event information from snippets
that contain sentences speaking about the same
event. H ¨urriyeto ˘glu et al. (2019) have defined that
an event has the following information classes (ex-
ample in Figure 1):
•Time, which indicates when the protest took
place,
•Facility name, which indicates in which facil-
ity the protest took place,
•Organizer, which indicates who organized the
protest,
• Participant, which indicates who participated
in the protest,
•Place, which indicates where the protest took
place in a more general area than the facility
(city, region, ...),
•Target, which indicates against whom or what
the protest took place,
•Trigger, which is a specific word or group of
words that indicate that a protest took place
(examples: protested, attack, ...),
Thus, not all the snippets contain all the classes,
and they can contain several times the same classes.
Each information can be composed of one or sev-
eral adjacent words. Each snippet contains infor-
mation related to one and only one event.
As the data is already separated into groups
of sentences related to the same event, our ap-
proach consists of considering a task of named
entity recognition with the aforementioned classes.
Multilingual BERT has already been used for multi-
lingual named entity recognition and showed great
results compared to state-of-the-art models (Hakala
and Pyysalo, 2019).The data is in BIO format (Ramshaw and Mar-
cus, 1995), where each word has a B tag or an I tag
of a specific class or an O tag. The B tag means
beginning and marks the beginning of a new entity.
The tag I means inside, which has to be preceded
by another I tag or a B tag, and marks that the word
is inside an entity but not the first word of the entity.
Finally, the O-tag means outside, which means the
word is not part of an entity.
3 System overview
Our model is based on pre-trained multilingual
BERT (Devlin et al., 2019). This model has been
pretrained on multilingual Wikipedia texts. To bal-
ance the fact that the data is not equally distributed
between all the languages the authors used expo-
nential smoothed weighting to under-sample the
most present languages and over-sample the rarest
ones. This does not perfectly balance all the lan-
guages but it reduces the impact of low-resourced
languages.
The authors of the M-BERT paper shared the
weights of a pretrained model that we use to do fine-
tuning. Fine-tuning a model consists of taking an
already trained model on a specific task and using
this model as a starting point of the training for the
task of interest. This approach has reached state-
of-the-arts in numerous tasks. In the case of M-
BERT, the pre-training tasks are Masked Language
Modeling (MLM) and Next Sentence Prediction
(NSP).
To be able to learn our task, we add a dense layer
on top of the outputs of M-BERT and learn it during
the fine-tuning. All our models are fine-tuning all
the layers of M-BERT.
The implementation is the one from Hugging-
Face’s ‘transformers’ library (Wolf et al., 2020). To
train it on our data, the model is fine-tuned on each
sub-task.
3.1 Sub task 1 and 2
For sub-tasks 1 and 2, we approach these tasks
as binary sequence classification, as the goal is to
predict whether or not a document (sub-task 1) orsentence (sub-task 2) contains relevant information
about a protest event. Thus the size of the output
of the dense layer is 2. We then perform an argmax
on these values to predict a class. We use the base
parameters in HuggingFace’s ’transformers’ library.
The loss is a cross-entropy, the learning rate is
handled by an AdamW optimizer (Loshchilov and
Hutter, 2019) and the activation function is a gelu
(Hendrycks and Gimpel, 2016). We use a dropout
of 10% for the fully connected layers inside M-
BERT and the attention probabilities.
One of the issues with M-BERT is the limited
length of the input, as it can only take 512 tokens,
which are tokenized words. M-BERT uses the
wordpiece tokenizer (Wu et al., 2016). A token is
either a word if the tokenizer knows it, if it does not
it will separate it into several sub-tokens which are
known. For sub-task 1, as we are working with en-
tire documents, it can be frequent that a document
is longer than this limit and has to be broken down
into several sub-documents. To retain contexts in
each sub-documents we use an overlap of 150 to-
kens, which means between two sub-documents,
they will have 150 tokens in common. Our method
to output a class, in this case, is as follows:
• tokenize a document,
•if the tokenized document is longer than
the 512-tokens limit, create different sub-
documents with 150-tokens overlaps between
each sub-document,
• generate a prediction for each sub-document,
•average all the predictions from sub-
documents originated from the same docu-
ment,
• take the argmax of the final prediction.
3.2 Sub-task 4
For sub-task 4, our approach is based on word
classification where we predict a class for each
word of the documents.
One issue is that as words are tokenized and can
be transformed into several sub-tokens we have to
choose how to choose the prediction of a multi-
token word. Our approach is to take the prediction
of the first token composing a word as in Hakala
and Pyysalo (2019).
We also have to deal with the input size as some
documents are longer than the limit. In this case,we separate them into sub-documents with an over-
lap of 150. Our approach is:
• tokenize a document,
•if the tokenized document is longer than
the 512-tokens limit, create different sub-
documents with 150-tokens overlaps between
each sub-document,
• generate a prediction for each sub-document,
•reconstruct the entire document: take the first
and second sub-documents, average the pre-
diction for the same tokens (from the overlap),
keep the prediction for the others, then use
the same process with the obtained document
and the next sub-document. As the size of
each sequence is 512 and the overlap is only
150, no tokens can be in more than 2 different
sequences,
•take the argmax of the final prediction for each
word.
3.2.1 Soft macro-F1 loss
We used a soft macro-F1 loss (Lipton et al., 2014).
This loss is closer than categorical cross-entropy on
BIO labels to the metric used to evaluate systems
in the shared task. The main issue with F1 is its
non-differentiability, so it cannot be used as is but
must be modified to become differentiable. The F1
score is based on precision and recall, which in turn
are functions of the number of true positives, false
positives, and false negatives. These quantities are
usually defined as follows:
tp=X
i2tokens(pred (i)true (i))
fp=X
i2tokens(pred (i)(1true (i)))
fn=X
i2tokens((1pred (i))true (i))
With:
•tokens , the list of tokens in a document,
•true(i) , 0 if the true label of the token i is of
the negative class, 1 if the true label is of the
positive class
•pred(i) , 0 if the predicted label of the token i
is of the negative class, 1 if the predicted label
is of the positive classAs we use macro-F1 loss, we compute the F1
score for each class where the positive class is the
current class and negative any other class, e.g. if
the reference class is B-trigger, then true(i)=1 for
B-trigger and true(i)=0 for all other classes when
macro-averaging the F1.
We replace the binary function pred(i) by a func-
tion outputting the predicted probability of the to-
ken i to be of the positive class:
soft tp=X
i2tokens(proba (i)true (i))
soft fp=X
i2tokens(proba (i)(1true (i)))
soft fn=X
i2tokens((1proba (i))true (i))
With proba(i) outputting the probability of the
token i to be of the positive class, this probability is
the predicted probability resulting from the softmax
activation of the fine-tuning network.
Then we compute, in a similar fashion as a nor-
mal F1, the precision and recall using the soft defi-
nitions of the true positive, false positive, and false
negative. And finally we compute the F1 score with
the given precision and recall. As a loss function is
a criterion to be minimized whereas F1 is a score
that we would like to maximize, the final loss is
1F1.
3.2.2 Recommendation for improved stability
A known problem of Transformers-based models
is the training instability, especially with small
datasets (Dodge et al., 2020; Ruder, 2021). Dodge
et al. (2020) explain that two elements that have
much influence on the stability are the data order
and the initialization of the prediction layer, both
controlled by pseudo-random numbers generated
from a seed. To study the impact of these two el-
ements on the models’ stability, we freeze all the
randomness on the other parts of the models and
change only two different random seeds:
•the data order, i.e. the different batches and
their order. Between two runs the model will
see the same data during each epoch but the
batches will be different, as the batches are
built beforehand and do not change between
epochs,
•the initialization of the linear layer used to
predict the output of the model.Another recommendation to work with
Transformers-based models and small data made
by Mosbach et al. (2021) is to use smaller learning
rates but compensating with more epochs. We have
taken this into account during the hyper-parameter
search.
Ruder (2021) recommend using behavioral fine-
tuning to reduce fine-tuning instabilities. It is sup-
posed to be especially helpful to have a better ini-
tialization of the final prediction layer. It has also al-
ready been used on named entity recognition tasks
(Broscheit, 2019) and has shown that it has im-
proved results for a task with a very small training
dataset. Thus, to do so, we need a task with the
same number of classes, but much larger training
datasets. As we did not find such a task, we de-
cided to fine-tune our model on at least the different
languages we are working with, English, Spanish
and Portuguese. We used named entity recognition
datasets and kept only three classes in common in
all the datasets: person, organization, and location.
These three types of entities can be found in the
shared task.
To perform this test, the training has been done
like that:
•the first fine-tuning is done on the concatena-
tion of NER datasets in different languages,
once the training is finished we save all the
weights of the model,
•we load the weights of the previous model,
except for the weights of the final prediction
layer which are randomized with a given seed,
•we train the model on the dataset of the shared
task.
4 Experimental setup
4.1 Data
The dataset of the shared task is based on articles
from different newspapers in different languages.
More information about this dataset can be found
in (H ¨urriyeto ˘glu et al., 2021a)
For the final submissions of sub-tasks 1, 2, and 4
we divided the dataset given for training purposes
into two parts with 80% for training and 20% for
evaluation during the system training phase. We
then predicted the data given for testing purposes
during the shared task evaluation phase. The quan-
tity of data for each sub-task and language can be
found in Table 1. We can note that the majority ofSub-task English Spanish Portuguese
Sub-task 1 9,324 1,000 1,487
Sub-task 2 22,825 2,741 1,182
Sub-task 4 808 33 30
Table 1: Number of elements for each sub-task for each
language in the data given for training purposes. Docu-
ments for sub-task 1, sentences for sub-task 2, snippet
(group of sentences about one event) for sub-task 4.
Dataset Train Eval Test
CoNLL 2003 14,041 3,250 3,453
CoNLL 2002 8,324 1,916 1,518
HAREM 121 8 128
Table 2: Number of elements for each dataset used in
the behavioral fine-tuning in each split.
the data is in English. Spanish and Portuguese are
only a small part of the dataset.
For all the experiments made on sub-task 4, we
divided the dataset given for training purposes into
three parts with 60% for training, 20% for evaluat-
ing and 20% for testing.
To be able to do our approach of behavioral fine-
tuning, we needed some Named Entity Recognition
datasets in English, Spanish and Portuguese. For
English we used the CoNLL 2003 dataset (Tjong
Kim Sang and De Meulder, 2003), for Spanish the
Spanish part of the CoNLL 2002 dataset (Tjong
Kim Sang, 2002) and for Portuguese the HAREM
dataset (Santos et al., 2006). Each of these datasets
had already three different splits for training, devel-
opment and test. Information about their size can
be found in Table 2.
The dataset for Portuguese is pretty small com-
pared to the two others, but the impact of the size
can be interesting to study.
4.2 Hyper-parameter search
For sub-task 4, we did a hyper-parameter search
to optimize the results. We used Ray Tune (Liaw
et al., 2018) and the HyperOpt algorithm Bergstra
et al. (2013). We launched 30 different trainings,
all the information about the search space and the
hyper-parameters can be found in A.1. The goal is
to optimize the macro-F1 on the evaluation set.
Our goal was to find a set of hyper-parameters
that performs well to use always the same in the
following experiments. We also wanted to evaluate
the impacts of the hyper-parameters on the training.4.3 Behavioral fine-tuning
For the first part of the behavioral fine-tuning,
we trained an M-BERT model on the three NER
datasets for one epoch. We only learn for one epoch
for timing issues, as the learning on this datasets
takes several hours. We then fine-tune the resulting
models with the best set of hyper-parameters found
with the hyper-parameter search.
4.4 Stability
To study the stability of the model and the impact
of behavioral fine-tuning we made 6 sets of experi-
ments with 20 experiments in each set:
•normal fine-tuning with random data order
and frozen initialization of final layer,
•normal fine-tuning with frozen data order and
random initialization of final layer,
•normal fine-tuning with random data order
and random initialization of final layer,
•behavioral fine-tuning with random data order
and frozen initialization of final layer,
•behavioral fine-tuning with frozen data order
and random initialization of final layer,
•behavioral fine-tuning with random data order
and random initialization of final layer,
Once again it is important to note that what we
called behavioral fine-tuning is different from be-
havioral fine-tuning as proposed by Ruder (2021),
as we reset the final layer. Only the weights of all
the layers of M-BERT are modified.
For each set of experiments we will look at
the average of the macro-F1, as implemented in
Nakayama (2018), and the standard deviation of
the macro-F1 on the training dataset, on the evalua-
tion dataset, and on three different test datasets, one
for each language. Thus we will be able to assess
the importance of the instability, if our approach to
behavioral fine-tuning helps to mitigate it and if it
has similar results across the languages.
We can also note that in our implementation
the batches are not randomized. They are built
once before the learning phase and do not change,
neither in content nor order of passage, between
each epoch.Figure 2: (Top) Parallel coordinates plot of the 30 experiments on sub-task 4 during the hyper-parameter search in
function of the value of the hyper-parameters and the value of the F1 on the evaluation set. Each line represents an
experiment, and each column a specific hyper-parameter, except the last which is the value of the metric. (Bottom)
Same plot with the worst results removed to have a better view of the best results.
5 Results
5.1 Hyper-parameter search
The results of the hyper-parameter search can be
seen in Figure 2. On the top pictures which repre-
sent the 30 experiments, we can see that a specific
hyper-parameter seems to impact the worst results
(in blue). This parameter is the learning rate, we
can see it in the red box on the top image, all the
blue lines are at the bottom, which means these
experiments had a small learning rate. It seems
that we obtain the best results with a learning rate
around 5e-05 (0.00005), lower than 1e-06 seems to
give bad results.
We can then focus on the bottom picture, with
the same type of plot but with the worst results
removed. Another hyper-parameter that seems to
have an impact is the number of training epochs,
40 seems better than 20. We use a high number of
epochs as recommended by Mosbach et al. (2021)
to limit the instability. Beyond the learning rate and
number of epochs, it is then hard to find impactful
hyper-parameters.
Finally, the set of hyper-parameters that has been
selected is:• Adafactor: True
• Number of training epochs: 40
• Adam beta 2: 0.99
• Adam beta 1: 0.74
• Maximum gradient norm: 0.17
• Adam epsilon: 3e-08
• Learning rate: 5e-05
• Weight decay: 0.36
For the stability experiments, the number of
training epochs have been reduced to 20 for speed
purposes. For the first part of the behavioral fine-
tuning, the learning rate has been set to 1e-05 as
more data were available.
5.2 Behavioral fine-tuning
The results on the test dataset of each model after
one epoch of training can be found in Table 5.
We could not compare to state-of-the-art NER
models on these three datasets as we do not take all
the classes (classes such as MISC were removedData Init layer Train Eval Test EN Test ES Test PT
NRand Fix 86.11 (1.08) 69.34 (1.01) 71.80 (.85) 54.33 (3.43) 73.14 ( 1.96)
Fix Rand 86.88 (.53) 70.03 (.63) 71.68 ( .53)55.02 (3.28) 74.51 (2.41)
Rand Rand 86.63 (1.08) 69.56 (.97) 71.94 (.72) 54.73 (3.44) 74.08 (3.37)
BRand Fix 85.79 (.97) 69.32 (1.00) 71.60 (.54) 54.69 (2.99) 74.01 (2.92)
Fix Rand 86.20 (.55) 69.57 ( .51)71.80 (.58) 53.97 (3.90) 74.50 (2.67)
Rand Rand 86.11 (.87) 69.40 (.80) 71.85 (.73) 55.51 (2.82)74.97 (2.66)
Table 3: Average macro-F1 score, higher is better (standard deviation, lower is better) of the 20 experiments with
the specified setup. N means normal fine-tuning and B behavioral fine-tuning. Data means data order and Init layer
means initialization of the final layer. Rand means random, and fix refers to frozen.
English Spanish Portuguese Hindi
Sub-task 1 53.46 (84.55) 46.47 (77.27) 46.47 (84.00) 29.66 (78.77)
Sub-task 2 75.64 (85.32) 76.39 (88.61) 81.61 (88.47) /
Sub-task 4 69.96 (78.11) 56.64 (66.20) 61.87 (73.24) /
Table 4: Score of our final submissions for each sub-task, in parenthesis the score achieved by the best scoring
team on each sub-task.
Dataset Test macro-F1
CoNLL 2003 89.8
CoNLL 2002 86.1
HAREM 76.1
Table 5: Macro-F1 score of the NER task on the test
split of each dataset used in behavioral fine-tuning after
training the base M-BERT for 1 epoch.
before the learning phase). The metrics used on
these datasets are not by classes, so the comparison
cannot be made. However, the results are already
much better than what a random classifier would
output, thus the weights of the models should al-
ready be better than the weights of the base model.
5.3 Stability
The results of the different sets of experiments can
be found in Table 3. First, we can see that the dif-
ference between behavioral fine-tuning and normal
fine-tuning is not important enough to say one is
better than the other. We can also note that the
standard deviation is small for English, but not
negligible for Spanish and Portuguese.
5.4 Final submission
The results of the final submissions can be found
in Table 4. We can see that our results are lower
than the best results, especially for sub-task 1 with
a difference of between 30 to 50 macro-F1 score
depending on the language, whereas for sub-tasks2 and 4 the difference is close to 10 macro-F1 score
for all the languages.
6 Conclusion
6.1 Sub-task 1 and 2
As we can see in Table 4, our final results for sub-
task 1 are much lower than the best results, but for
sub-task 2 the difference is smaller. This is interest-
ing as the tasks are pretty similar, thus expected the
difference between our results and the best results
to be of the same magnitude.
One explanation could be our approach to han-
dle documents longer than the input of M-BERT.
We have chosen to take the average of the sub-
documents, but if one part of a document contains
an event the entire document does too. We may
have better results looking if one sub-document at
least is considered as having an event.
It is then hard to compare to other models as we
have chosen to use one model for all the languages
and we do not know the other approaches.
6.2 Sub-task 4
For sub-task 4 we have interesting results for all
the languages, even for Spanish and Portuguese,
as we were not sure that we could learn this task
in a supervised fashion with the amount of data
available. In a further study, we could compare our
results with results obtained by fine-tuning mono-
lingual models, where we fine-tune one model for
each language with only the data of one language.This could show the impact of having data if using
a multilingual model instead of several monolin-
gual models improves or not the results. We do not
expect good results for Spanish and Portuguese as
the training dataset is pretty limited. The results
seem to comfort the claim of (Pires et al., 2019)
that M-BERT works well for few-shot learning on
other languages.
The other question for sub-task 4 was about in-
stability. In Table 3 we can see that the instability is
way more pronounced for Spanish and Portuguese.
It seems logical as we have fewer data available
in Spanish and Portuguese than in English. The
standard deviation for Spanish and Portuguese is
large and can have a real impact on the final re-
sults. Finding good seeds could help to improve
the results for Spanish and Portuguese.
Furthermore, our approach of behavioral fine-
tuning did not help to reduce the instabilities. It
was expected that one of the sources of the insta-
bility is the initialization of the prediction, and in
our approach, the initialization of this layer is still
random. In our approach, we only fine-tune the
weights of M-BERT. This does not seem to work
and reinforces the advice of Ruder (2021) that us-
ing behavioral fine-tuning is more useful for having
a good initialization of the final prediction layer.
On the two sources of randomness we studied,
data order seems the most impactful for English,
where we have more data. Nonetheless, for Span-
ish and Portuguese, the two sources have a large
impact. In a further study, we could see how the
quantity of data helps to decrease the impact of
these sources of instabilities.
For the final submissions, the macro-F1 score
for English and Portuguese is beneath the average
macro-F1 score we found during our development
phases. This could be due to bad seeds for random-
ness or because the splits are different. We did not
try to find the best-performing seeds for the final
submissions.
Acknowledgments
We thank Damien Fourrure, Arnaud Jacques, Guil-
laume Stempfel and our anonymous reviewers for
their helpful comments.
References
James Bergstra, Daniel Yamins, and David Cox. 2013.
Making a science of model search: Hyperparameteroptimization in hundreds of dimensions for vision ar-
chitectures. In International conference on machine
learning , pages 115–123. PMLR.
Samuel Broscheit. 2019. Investigating entity knowl-
edge in BERT with simple neural end-to-end en-
tity linking. In Proceedings of the 23rd Confer-
ence on Computational Natural Language Learning
(CoNLL) , pages 677–685, Hong Kong, China. Asso-
ciation for Computational Linguistics.
Jacob Devlin, Ming-Wei Chang, Kenton Lee, and
Kristina Toutanova. 2019. Bert: Pre-training of
deep bidirectional transformers for language under-
standing. In Proceedings of the 2019 Conference of
the North American Chapter of the Association for
Computational Linguistics: Human Language Tech-
nologies, Volume 1 (Long and Short Papers) , pages
4171–4186.
Jesse Dodge, Gabriel Ilharco, Roy Schwartz, Ali
Farhadi, Hannaneh Hajishirzi, and Noah Smith.
2020. Fine-tuning pretrained language models:
Weight initializations, data orders, and early stop-
ping. arXiv preprint arXiv:2002.06305 .
Kai Hakala and Sampo Pyysalo. 2019. Biomedical
named entity recognition with multilingual BERT.
InProceedings of The 5th Workshop on BioNLP
Open Shared Tasks , pages 56–61, Hong Kong,
China. Association for Computational Linguistics.
Dan Hendrycks and Kevin Gimpel. 2016. Gaus-
sian error linear units (gelus). arXiv preprint
arXiv:1606.08415 .
Ali H ¨urriyeto ˘glu, Osman Mutlu, Farhana Ferdousi
Liza, Erdem Y ¨or¨uk, Ritesh Kumar, and Shyam
Ratan. 2021a. Multilingual protest news detection
- shared task 1, case 2021. In Proceedings of the 4th
Workshop on Challenges and Applications of Auto-
mated Extraction of Socio-political Events from Text
(CASE 2021) , online. Association for Computational
Linguistics (ACL).
Ali H ¨urriyeto ˘glu, Hristo Tanev, Vanni Zavarella, Jakub
Piskorski, Reyyan Yeniterzi, and Erdem Y ¨or¨uk.
2021b. Challenges and applications of automated
extraction of socio-political events from text (case
2021): Workshop and shared task report. In
Proceedings of the 4th Workshop on Challenges
and Applications of Automated Extraction of Socio-
political Events from Text (CASE 2021) , online. As-
sociation for Computational Linguistics (ACL).
Ali H ¨urriyeto ˘glu, Erdem Y ¨or¨uk, Deniz Y ¨uret, C ¸ a ˘grı
Yoltar, Burak G ¨urel, Fırat Durus ¸an, Osman Mutlu,
and Arda Akdemir. 2019. Overview of clef 2019
lab protestnews: Extracting protests from news
in a cross-context setting. In Experimental IR
Meets Multilinguality, Multimodality, and Interac-
tion, pages 425–432, Cham. Springer International
Publishing.Ali H ¨urriyeto ˘glu, Vanni Zavarella, Hristo Tanev, Er-
dem Y ¨or¨uk, Ali Safaya, and Osman Mutlu. 2020.
Automated extraction of socio-political events from
news (AESPEN): Workshop and shared task report.
InProceedings of the Workshop on Automated Ex-
traction of Socio-political Events from News 2020 ,
pages 1–6, Marseille, France. European Language
Resources Association (ELRA).
Richard Liaw, Eric Liang, Robert Nishihara, Philipp
Moritz, Joseph E Gonzalez, and Ion Stoica.
2018. Tune: A research platform for distributed
model selection and training. arXiv preprint
arXiv:1807.05118 .
Zachary C Lipton, Charles Elkan, and Balakrishnan
Naryanaswamy. 2014. Optimal thresholding of clas-
sifiers to maximize f1 measure. In Joint European
Conference on Machine Learning and Knowledge
Discovery in Databases , pages 225–239. Springer.
Ilya Loshchilov and Frank Hutter. 2019. Decoupled
weight decay regularization. In International Con-
ference on Learning Representations .
Marius Mosbach, Maksym Andriushchenko, and Diet-
rich Klakow. 2021. On the stability of fine-tuning
fbertg: Misconceptions, explanations, and strong
baselines. In International Conference on Learning
Representations .
Hiroki Nakayama. 2018. seqeval: A python framework
for sequence labeling evaluation. Software available
from https://github.com/chakki-works/seqeval.
Telmo Pires, Eva Schlinger, and Dan Garrette. 2019.
How multilingual is multilingual bert? In Proceed-
ings of the 57th Annual Meeting of the Association
for Computational Linguistics , pages 4996–5001.
Lance Ramshaw and Mitch Marcus. 1995. Text chunk-
ing using transformation-based learning. In Third
Workshop on Very Large Corpora .
Sebastian Ruder. 2021. Recent Advances in Lan-
guage Model Fine-tuning. http://ruder.io/
recent-advances-lm-fine-tuning .
Diana Santos, Nuno Seco, Nuno Cardoso, and Rui
Vilela. 2006. Harem: An advanced ner evalua-
tion contest for portuguese. In quot; In Nicoletta
Calzolari; Khalid Choukri; Aldo Gangemi; Bente
Maegaard; Joseph Mariani; Jan Odjik; Daniel
Tapias (ed) Proceedings of the 5 th International
Conference on Language Resources and Evaluation
(LREC’2006)(Genoa Italy 22-28 May 2006) .
Erik F. Tjong Kim Sang. 2002. Introduction to the
CoNLL-2002 shared task: Language-independent
named entity recognition. In COLING-02: The
6th Conference on Natural Language Learning 2002
(CoNLL-2002) .
Erik F. Tjong Kim Sang and Fien De Meulder.
2003. Introduction to the CoNLL-2003 shared task:
Language-independent named entity recognition. InProceedings of the Seventh Conference on Natu-
ral Language Learning at HLT-NAACL 2003 , pages
142–147.
Thomas Wolf, Lysandre Debut, Victor Sanh, Julien
Chaumond, Clement Delangue, Anthony Moi, Pier-
ric Cistac, Tim Rault, R ´emi Louf, Morgan Funtow-
icz, Joe Davison, Sam Shleifer, Patrick von Platen,
Clara Ma, Yacine Jernite, Julien Plu, Canwen Xu,
Teven Le Scao, Sylvain Gugger, Mariama Drame,
Quentin Lhoest, and Alexander M. Rush. 2020.
Transformers: State-of-the-art natural language pro-
cessing. In Proceedings of the 2020 Conference on
Empirical Methods in Natural Language Processing:
System Demonstrations , pages 38–45, Online. Asso-
ciation for Computational Linguistics.
Yonghui Wu, Mike Schuster, Zhifeng Chen, Quoc V .
Le, Mohammad Norouzi, Wolfgang Macherey,
Maxim Krikun, Yuan Cao, Qin Gao, Klaus
Macherey, Jeff Klingner, Apurva Shah, Melvin John-
son, Xiaobing Liu, Łukasz Kaiser, Stephan Gouws,
Yoshikiyo Kato, Taku Kudo, Hideto Kazawa, Keith
Stevens, George Kurian, Nishant Patil, Wei Wang,
Cliff Young, Jason Smith, Jason Riesa, Alex Rud-
nick, Oriol Vinyals, Greg Corrado, Macduff Hughes,
and Jeffrey Dean. 2016. Google’s neural machine
translation system: Bridging the gap between human
and machine translation. CoRR , abs/1609.08144.
A Appendix
A.1 Hyper-parameter search
The space search for our hyper-parameter search
was:
•Number of training epochs: value in [20, 25,
30, 40],
•Weight decay: uniform distribution between
0.001 and 1,
•Learning rate: value in [1e-5, 2e-5, 3e-5, 4e-5,
5e-5, 6e-5, 2e-7, 1e-7, 3e-7, 2e-8],
• Adafactor: value in ”True”, ”False”,
•Adam beta 1: uniform distribution between 0
and 1,
•Adam beta 2: uniform distribution between 0
and 1,
•Epsilon: value in [1e-8, 2e-8, 3e-8, 1e-9, 2e-9,
3e-10],
•Maximum gradient norm: uniform distribu-
tion between 0 and 1.For the HyperOpt algorithm we used two set
of hyper-parameters to help finding a good sub-
space. We maximized the macro-F1 on the evalu-
ation dataset, and set the number of initial points
before starting the algorithm to 5.