elmadany commited on
Commit
f1aab30
·
verified ·
1 Parent(s): cc8a712

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md CHANGED
@@ -10,3 +10,36 @@
10
  - text: "ومن أشعاره الفصيحة ما يأتي:\nدع الأيام تفعل ما تشاء ** وطب نفسا إذا حكم القضاء\nولا تجزع لنازلة الليالي ** فما لحوادث الدنيا بقاء\nوكن رجلا على الأهوال جلدا ** وشيمتك"
11
 
12
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  - text: "ومن أشعاره الفصيحة ما يأتي:\nدع الأيام تفعل ما تشاء ** وطب نفسا إذا حكم القضاء\nولا تجزع لنازلة الليالي ** فما لحوادث الدنيا بقاء\nوكن رجلا على الأهوال جلدا ** وشيمتك"
11
 
12
  ---
13
+ # Jasmine-350M
14
+ # JASMINE: Arabic GPT Models for Few-Shot Learning
15
+
16
+
17
+ This is the repository accompanying our EMNLP2023 paper [JASMINE: Arabic GPT Models for Few-Shot Learning](https://aclanthology.org/2023.emnlp-main.1040/). JASMINE is a suite of powerful Arabic autoregressive Transformer language models ranging in size between 300 million-6.7 billion parameters pretrained on a large and diverse dataset ( 235 GB of text).
18
+
19
+ # BibTex
20
+
21
+ If you use Jasmine models for your scientific publication, or if you find the resources in this repository useful, please cite our paper as follows (to be updated):
22
+ ```bibtex
23
+ @inproceedings{billah-nagoudi-etal-2023-jasmine,
24
+ title = "{JASMINE}: {A}rabic {GPT} Models for Few-Shot Learning",
25
+ author = "Billah Nagoudi, El Moatez and
26
+ Abdul-Mageed, Muhammad and
27
+ Elmadany, AbdelRahim and
28
+ Inciarte, Alcides and
29
+ Islam Khondaker, Md Tawkat",
30
+ editor = "Bouamor, Houda and
31
+ Pino, Juan and
32
+ Bali, Kalika",
33
+ booktitle = "Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing",
34
+ month = dec,
35
+ year = "2023",
36
+ address = "Singapore",
37
+ publisher = "Association for Computational Linguistics",
38
+ url = "https://aclanthology.org/2023.emnlp-main.1040",
39
+ doi = "10.18653/v1/2023.emnlp-main.1040",
40
+ pages = "16721--16744",
41
+ abstract = "Scholarship on generative pretraining (GPT) remains acutely Anglocentric, leaving serious gaps in our understanding of the whole class of autoregressive models. For example, we have little knowledge about the potential of these models and their societal impacts in diverse linguistic and cultural settings. We alleviate this issue for Arabic, a wide collection of languages and dialectal varieties with more than 400 million population, by introducing JASMINE. JASMINE is a suite of powerful Arabic autoregressive Transformer language models ranging in size between 300 million-6.7 billion parameters pretrained on a large and diverse dataset ( 235 GB of text). We also carefully design and release a comprehensive benchmark for both automated and human evaluation of Arabic autoregressive models, with coverage of potential social biases, harms, and toxicity. Using our novel benchmark, we evaluate JASMINE extensively showing powerful performance intrinsically as well as in few-shot learning on a wide range of NLP tasks. We aim to responsibly release our models and evaluation benchmark with interested researchers, along with code for experimenting with them.",
42
+ }
43
+
44
+ ## Acknowledgments
45
+ We gratefully acknowledge support from the Natural Sciences and Engineering Research Council of Canada, the Social Sciences and Humanities Research Council of Canada, Canadian Foundation for Innovation, [ComputeCanada](www.computecanada.ca) and [UBC ARC-Sockeye](https://doi.org/10.14288/SOCKEYE). We also thank the [Google TensorFlow Research Cloud (TFRC)](https://www.tensorflow.org/tfrc) program for providing us with free TPU access.