GGUF
Russian
conversational
IlyaGusev commited on
Commit
2e9a739
·
2 Parent(s): ac134dd 07c2d69

Merge branch 'main' of https://huggingface.co/IlyaGusev/saiga_yandexgpt_8b_gguf into main

Browse files
Files changed (1) hide show
  1. README.md +33 -0
README.md ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ datasets:
3
+ - IlyaGusev/saiga_scored
4
+ - IlyaGusev/saiga_preferences
5
+ language:
6
+ - ru
7
+ inference: false
8
+ license: other
9
+ license_name: yandexgpt-5-lite-8b-pretrain
10
+ license_link: LICENSE
11
+ ---
12
+
13
+ Llama.cpp compatible versions of an original [8B model](https://huggingface.co/IlyaGusev/saiga_yandexgpt_8b).
14
+
15
+ Download one of the versions, for example `saiga_yandexgpt_8b.Q4_K_M.gguf`.
16
+ ```
17
+ wget https://huggingface.co/IlyaGusev/saiga_yandexgpt_8b_gguf/resolve/main/saiga_yandexgpt_8b.Q4_K_M.gguf
18
+ ```
19
+
20
+ Download [interact_gguf.py](https://raw.githubusercontent.com/IlyaGusev/saiga/refs/heads/main/scripts/interact_gguf.py)
21
+ ```
22
+ https://raw.githubusercontent.com/IlyaGusev/saiga/refs/heads/main/scripts/interact_gguf.py
23
+ ```
24
+
25
+ How to run:
26
+ ```
27
+ pip install llama-cpp-python fire
28
+
29
+ python3 interact_gguf.py saiga_yandexgpt_8b.Q4_K_M.gguf
30
+ ```
31
+
32
+ System requirements:
33
+ * 15GB RAM for q8_0 and less for smaller quantizations