File size: 580 Bytes
7411d9c
c8a4e29
37348d2
 
8dc9a4e
 
 
1e1a2ca
 
 
1
2
3
4
5
6
7
8
9
10
---
license: agpl-3.0
language:
- sr
---


Model is developed in support of the University of Belgrade doctoral dissertation "Composite pseudogrammars based on parallel language models of Serbian" by Mihailo Škorić.

This small gpt-2 model was trained on several corpora for Serbian, including ["The corpus of Contemporary Serbian"](https://drive.google.com/file/d/1wRgoWer6YULGCXR0zWOl1fVA6VIe1DOR), [SrpELTeC](https://drive.google.com/file/d/1RtBXyw5Cdh6y_cqbJoMlYhSwNFydBRUv) and WikiKorpus by [JeRTeh – Society for Language Resources and Technologies](https://jerteh.rs/).