Update model card
Browse files
README.md
CHANGED
@@ -1,3 +1,60 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language: multilingual
|
3 |
+
widget:
|
4 |
+
- text: 🤗🤗🤗<mask>
|
5 |
+
- text: 🔥The goal of life is <mask> . 🔥
|
6 |
+
- text: Il segreto della vita è l’<mask> . ❤️
|
7 |
+
- text: Hasta <mask> 👋!
|
8 |
license: mit
|
9 |
---
|
10 |
+
|
11 |
+
|
12 |
+
# Twitter-XLM-Roberta-large
|
13 |
+
This is a XLM-T large language model specialised on Twitter.
|
14 |
+
The base model was the multilingual XLM-R and the model was then re-trained on tweets from many different languages until December 2022.
|
15 |
+
|
16 |
+
To evaluate this and other LMs on Twitter-specific data, please refer to the [XLM-T main repository](https://github.com/cardiffnlp/xlm-t).
|
17 |
+
|
18 |
+
Finally, this model is fully compatible with the [TweetNLP library](https://github.com/cardiffnlp/tweetnlp)
|
19 |
+
|
20 |
+
|
21 |
+
```
|
22 |
+
|
23 |
+
### BibTeX entry and citation info
|
24 |
+
|
25 |
+
More information in the reference papers about [multilingual language models on Twitter](https://aclanthology.org/2022.lrec-1.27/) and [time-specific models](https://aclanthology.org/2022.acl-demo.25/).
|
26 |
+
Please cite the relevant reference papers if you use this model.
|
27 |
+
|
28 |
+
```bibtex
|
29 |
+
@inproceedings{barbieri-etal-2022-xlm,
|
30 |
+
title = "{XLM}-{T}: Multilingual Language Models in {T}witter for Sentiment Analysis and Beyond",
|
31 |
+
author = "Barbieri, Francesco and
|
32 |
+
Espinosa Anke, Luis and
|
33 |
+
Camacho-Collados, Jose",
|
34 |
+
booktitle = "Proceedings of the Thirteenth Language Resources and Evaluation Conference",
|
35 |
+
month = jun,
|
36 |
+
year = "2022",
|
37 |
+
address = "Marseille, France",
|
38 |
+
publisher = "European Language Resources Association",
|
39 |
+
url = "https://aclanthology.org/2022.lrec-1.27",
|
40 |
+
pages = "258--266",
|
41 |
+
abstract = "Language models are ubiquitous in current NLP, and their multilingual capacity has recently attracted considerable attention. However, current analyses have almost exclusively focused on (multilingual variants of) standard benchmarks, and have relied on clean pre-training and task-specific corpora as multilingual signals. In this paper, we introduce XLM-T, a model to train and evaluate multilingual language models in Twitter. In this paper we provide: (1) a new strong multilingual baseline consisting of an XLM-R (Conneau et al. 2020) model pre-trained on millions of tweets in over thirty languages, alongside starter code to subsequently fine-tune on a target task; and (2) a set of unified sentiment analysis Twitter datasets in eight different languages and a XLM-T model trained on this dataset.",
|
42 |
+
}
|
43 |
+
|
44 |
+
@inproceedings{loureiro-etal-2022-timelms,
|
45 |
+
title = "{T}ime{LM}s: Diachronic Language Models from {T}witter",
|
46 |
+
author = "Loureiro, Daniel and
|
47 |
+
Barbieri, Francesco and
|
48 |
+
Neves, Leonardo and
|
49 |
+
Espinosa Anke, Luis and
|
50 |
+
Camacho-collados, Jose",
|
51 |
+
booktitle = "Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: System Demonstrations",
|
52 |
+
month = may,
|
53 |
+
year = "2022",
|
54 |
+
address = "Dublin, Ireland",
|
55 |
+
publisher = "Association for Computational Linguistics",
|
56 |
+
url = "https://aclanthology.org/2022.acl-demo.25",
|
57 |
+
doi = "10.18653/v1/2022.acl-demo.25",
|
58 |
+
pages = "251--260",
|
59 |
+
abstract = "Despite its importance, the time variable has been largely neglected in the NLP and language model literature. In this paper, we present TimeLMs, a set of language models specialized on diachronic Twitter data. We show that a continual learning strategy contributes to enhancing Twitter-based language models{'} capacity to deal with future and out-of-distribution tweets, while making them competitive with standardized and more monolithic benchmarks. We also perform a number of qualitative analyses showing how they cope with trends and peaks in activity involving specific named entities or concept drift. TimeLMs is available at github.com/cardiffnlp/timelms.",
|
60 |
+
}
|