Update README.md
Browse files
README.md
CHANGED
@@ -5,11 +5,11 @@ language:
|
|
5 |
pipeline_tag: text-generation
|
6 |
---
|
7 |
|
8 |
-
# Model Card for
|
9 |
|
10 |
<!-- Provide a quick summary of what the model is/does. -->
|
11 |
|
12 |
-
This model is pre-trained with the causal language modelling objective on a private web scraped
|
13 |
|
14 |
The dataset is cleaned and balanced with a specialized procedure to avoid cultural, political, racial and other biases. The procedure is described in the paper dedicated to this model- coming soon!
|
15 |
|
@@ -18,7 +18,7 @@ The dataset is cleaned and balanced with a specialized procedure to avoid cultur
|
|
18 |
|
19 |
### Model Description
|
20 |
|
21 |
-
|
22 |
|
23 |
|
24 |
|
@@ -39,7 +39,8 @@ The dataset is cleaned and balanced with a specialized procedure to avoid cultur
|
|
39 |
|
40 |
## Uses
|
41 |
|
42 |
-
|
|
|
43 |
|
44 |
### Direct Use
|
45 |
|
@@ -75,7 +76,17 @@ Users (both direct and downstream) should be made aware of the risks, biases and
|
|
75 |
|
76 |
Use the code below to get started with the model.
|
77 |
|
78 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
79 |
|
80 |
## Training Details
|
81 |
|
|
|
5 |
pipeline_tag: text-generation
|
6 |
---
|
7 |
|
8 |
+
# Model Card for GPT-WEB-BG
|
9 |
|
10 |
<!-- Provide a quick summary of what the model is/does. -->
|
11 |
|
12 |
+
This model is pre-trained with the causal language modelling objective on a private dataset with web scraped content created at the Bulgarian Academy of Sciences under the [ClaDa-BG Project](https://clada-bg.eu/en/).
|
13 |
|
14 |
The dataset is cleaned and balanced with a specialized procedure to avoid cultural, political, racial and other biases. The procedure is described in the paper dedicated to this model- coming soon!
|
15 |
|
|
|
18 |
|
19 |
### Model Description
|
20 |
|
21 |
+
The model is the first from a series of Large Languege Models for Bulgarian.
|
22 |
|
23 |
|
24 |
|
|
|
39 |
|
40 |
## Uses
|
41 |
|
42 |
+
The model is trained on the causal language modeling objective and can be used to generate content based on textual input. It can be further finetuned for specific NLP tasks in the online media domain such as Event Extraction, Relation Extracation, Named Entity Recognition, etc.
|
43 |
+
This model is intended for use from researchers and practitioners in the NLP field.
|
44 |
|
45 |
### Direct Use
|
46 |
|
|
|
76 |
|
77 |
Use the code below to get started with the model.
|
78 |
|
79 |
+
```python
|
80 |
+
from transformers import pipeline, set_seed
|
81 |
+
gpt_web_bg = pipeline('text-generation', model='/usmiva/gpt_web_bg', max_length=50, num_beams=3, temperature=0.8)
|
82 |
+
set_seed(42)
|
83 |
+
```
|
84 |
+
|
85 |
+
```python
|
86 |
+
gpt_web_bg("По професия той е ")
|
87 |
+
|
88 |
+
```
|
89 |
+
[{'generated_text': 'По професия той е строителен работник, който е �'}]
|
90 |
|
91 |
## Training Details
|
92 |
|