|
--- |
|
inference: false |
|
language: |
|
- bg |
|
license: mit |
|
datasets: |
|
- oscar |
|
- chitanka |
|
- wikipedia |
|
tags: |
|
- torch |
|
--- |
|
|
|
# ROBERTA BASE (cased) trained on private Bulgarian sentiment-analysis dataset |
|
This is a Multilingual Roberta model. |
|
|
|
This model is cased: it does make a difference between bulgarian and Bulgarian. |
|
|
|
### How to use |
|
|
|
Here is how to use this model in PyTorch: |
|
|
|
```python |
|
>>> import torch |
|
>>> from transformers import AutoModel, AutoTokenizer |
|
>>> |
|
>>> model_id = "rmihaylov/roberta-base-sentiment-bg" |
|
>>> model = AutoModel.from_pretrained(model_id, trust_remote_code=True) |
|
>>> tokenizer = AutoTokenizer.from_pretrained(model_id) |
|
>>> |
|
>>> inputs = tokenizer.batch_encode_plus(['Това е умно.', 'Това е тъпо.'], return_tensors='pt') |
|
>>> outputs = model(**inputs) |
|
>>> torch.softmax(outputs, dim=1).tolist() |
|
|
|
[[0.0004746630438603461, 0.9995253086090088], |
|
[0.9986956715583801, 0.0013043134240433574]] |
|
``` |
|
|