t5-super-tiny-standard-bahasa-cased
Pretrained T5 super-tiny standard language model for Malay.
Pretraining Corpus
t5-super-tiny-standard-bahasa-cased
model was pretrained on multiple tasks. Below is list of tasks we trained on,
- Language masking task on bahasa news, bahasa Wikipedia, bahasa Academia.edu, bahasa parliament and translated The Pile.
- News title prediction on bahasa news.
- Next sentence prediction on bahasa news, bahasa Wikipedia, bahasa Academia.edu, bahasa parliament and translated The Pile.
- Translated QA Natural.
- Text Similarity task on translated SNLI and translated MNLI.
- EN-MS translation.
- MS-EN translation.
- Abstractive Summarization.
- Knowledge Graph triples generation.
- Paraphrase.
Preparing steps can reproduce at https://github.com/huseinzol05/malaya/tree/master/pretrained-model/t5/prepare
Pretraining details
- This model was trained using Google T5 repository https://github.com/google-research/text-to-text-transfer-transformer, on v3-8 TPU.
- All steps can reproduce from here, https://github.com/huseinzol05/Malaya/tree/master/pretrained-model/t5
Load Pretrained Model
You can use this model by installing torch
or tensorflow
and Huggingface library transformers
. And you can use it directly by initializing it like this:
from transformers import T5Tokenizer, T5Model
model = T5Model.from_pretrained('malay-huggingface/t5-super-tiny-bahasa-cased')
tokenizer = T5Tokenizer.from_pretrained('malay-huggingface/t5-super-tiny-bahasa-cased')
Example using T5ForConditionalGeneration
from transformers import T5Tokenizer, T5ForConditionalGeneration
tokenizer = T5Tokenizer.from_pretrained('malay-huggingface/t5-super-tiny-bahasa-cased')
model = T5ForConditionalGeneration.from_pretrained('malay-huggingface/t5-super-tiny-bahasa-cased')
input_ids = tokenizer.encode('soalan: siapakah perdana menteri malaysia?', return_tensors = 'pt')
outputs = model.generate(input_ids)
print(tokenizer.decode(outputs[0]))
Output is,
'Mahathir Mohamad'
Supported prefix
soalan: {string}
, trained using Natural QA.ringkasan: {string}
, for abstractive summarization.tajuk: {string}
, for abstractive title.parafrasa: {string}
, for abstractive paraphrase.terjemah Inggeris ke Melayu: {string}
, for EN-MS translation.terjemah Melayu ke Inggeris: {string}
, for MS-EN translation.grafik pengetahuan: {string}
, for MS text to EN Knowledge Graph triples format.ayat1: {string1} ayat2: {string2}
, semantic similarity.
- Downloads last month
- 4
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.