|
--- |
|
license: mit |
|
language: |
|
- ru |
|
- en |
|
pipeline_tag: text-generation |
|
inference: false |
|
tags: |
|
- gpt3 |
|
- transformers |
|
- pytorch |
|
--- |
|
This is a generative model converted to fp16 format based on [ai-forever/ruGPT-3.5-13B](https://huggingface.co/ai-forever/ruGPT-3.5-13B) |
|
|
|
## Examples of usage |
|
|
|
```python |
|
from transformers import AutoTokenizer |
|
from auto_gptq import AutoGPTQForCausalLM |
|
|
|
model = AutoGPTQForCausalLM.from_quantized('Gaivoronsky/ruGPT-3.5-13B-8bit', device="cuda:0", use_triton=False) |
|
tokenizer = AutoTokenizer.from_pretrained('Gaivoronsky/ruGPT-3.5-13B-8bit') |
|
|
|
request = "Человек: Сколько весит жираф? Помощник: " |
|
encoded_input = tokenizer(request, return_tensors='pt', \ |
|
add_special_tokens=False).to('cuda') |
|
output = model.generate( |
|
**encoded_input, |
|
num_beams=2, |
|
do_sample=True, |
|
max_new_tokens=100 |
|
) |
|
print(tokenizer.decode(output[0], skip_special_tokens=True)) |
|
``` |