|
--- |
|
license: bigscience-bloom-rail-1.0 |
|
language: |
|
- tr |
|
--- |
|
Base Model: |
|
https://huggingface.co/bigscience/bloomz-7b1 |
|
|
|
--- |
|
Model fine-tuned on a real news dataset and optimized for neural news generation. |
|
|
|
Note: Turkish was not in pretraining. |
|
|
|
```python |
|
from transformers import AutoModelForSequenceClassification, AutoTokenizer, pipeline |
|
|
|
# Load model and tokenizer |
|
tokenizer = AutoTokenizer.from_pretrained('bigscience/bloomz') |
|
model = AutoModelForSequenceClassification.from_pretrained('tum-nlp/neural-news-generator-bloomz-7b1-tr') |
|
|
|
# Create the pipeline for neural news generation and set the repetition penalty >1.1 to punish repetition. |
|
generator = pipeline('text-generation', |
|
model=model, |
|
tokenizer=tokenizer, |
|
repetition_penalty=1.2) |
|
|
|
# Define the prompt |
|
prompt = "Başlık: Madde madde Türkiye, İsveç'in NATO mutabakatı Metin: Türkiye, İsveç ve NATO'nun Litvanya'nın başkenti Vilnius'ta mutabakatla biten üçlü görüşmesi [EOP]" |
|
|
|
# Generate |
|
generator(prompt, max_length=1000, num_return_sequences=1) |
|
|
|
``` |
|
|
|
Trained on 6k datapoints (including all splits) from: |
|
https://huggingface.co/datasets/batubayk/TR-News/tree/main |
|
|