kuotient commited on
Commit
f0c44a5
β€’
1 Parent(s): 2de9758

Upload 2 files

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +80 -0
  3. ggml-model-q4_k_m.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ ggml-model-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - ko
4
+
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ license: cc-by-nc-4.0
8
+ ---
9
+
10
+ # **Synatra-7B-v0.3-RP🐧**
11
+ ![Synatra-7B-v0.3-RP](./Synatra.png)
12
+
13
+ ## Support Me
14
+ μ‹œλ‚˜νŠΈλΌλŠ” 개인 ν”„λ‘œμ νŠΈλ‘œ, 1인의 μžμ›μœΌλ‘œ 개발되고 μžˆμŠ΅λ‹ˆλ‹€. λͺ¨λΈμ΄ λ§ˆμŒμ— λ“œμ…¨λ‹€λ©΄ μ•½κ°„μ˜ 연ꡬ비 지원은 μ–΄λ–¨κΉŒμš”?
15
+ [<img src="https://cdn.buymeacoffee.com/buttons/default-orange.png" alt="Buy me a Coffee" width="217" height="50">](https://www.buymeacoffee.com/mwell)
16
+
17
+ Wanna be a sponser? Contact me on Telegram **AlzarTakkarsen**
18
+
19
+ # **License**
20
+
21
+ This model is strictly [*non-commercial*](https://creativecommons.org/licenses/by-nc/4.0/) (**cc-by-nc-4.0**) use only.
22
+ The "Model" is completely free (ie. base model, derivates, merges/mixes) to use for non-commercial purposes as long as the the included **cc-by-nc-4.0** license in any parent repository, and the non-commercial use statute remains, regardless of other models' licences.
23
+ The licence can be changed after new model released. If you are to use this model for commercial purpose, Contact me.
24
+
25
+ # **Model Details**
26
+ **Base Model**
27
+ [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1)
28
+
29
+ **Trained On**
30
+ A6000 48GB * 8
31
+
32
+ **Instruction format**
33
+
34
+ It follows [ChatML](https://github.com/openai/openai-python/blob/main/chatml.md) format.
35
+
36
+ **TODO**
37
+
38
+ - ~~``RP 기반 νŠœλ‹ λͺ¨λΈ μ œμž‘``~~ βœ…
39
+ - ~~``데이터셋 μ •μ œ``~~ βœ…
40
+ - μ–Έμ–΄ 이해λŠ₯λ ₯ κ°œμ„ 
41
+ - ~~``상식 보완``~~ βœ…
42
+ - ν† ν¬λ‚˜μ΄μ € λ³€κ²½
43
+
44
+
45
+ # **Model Benchmark**
46
+
47
+ ## Ko-LLM-Leaderboard
48
+
49
+ On Benchmarking...
50
+
51
+ # **Implementation Code**
52
+
53
+ Since, chat_template already contains insturction format above.
54
+ You can use the code below.
55
+
56
+ ```python
57
+ from transformers import AutoModelForCausalLM, AutoTokenizer
58
+
59
+ device = "cuda" # the device to load the model onto
60
+
61
+ model = AutoModelForCausalLM.from_pretrained("maywell/Synatra-7B-v0.3-RP")
62
+ tokenizer = AutoTokenizer.from_pretrained("maywell/Synatra-7B-v0.3-RP")
63
+
64
+ messages = [
65
+ {"role": "user", "content": "λ°”λ‚˜λ‚˜λŠ” μ›λž˜ ν•˜μ–€μƒ‰μ΄μ•Ό?"},
66
+ ]
67
+
68
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
69
+
70
+ model_inputs = encodeds.to(device)
71
+ model.to(device)
72
+
73
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
74
+ decoded = tokenizer.batch_decode(generated_ids)
75
+ print(decoded[0])
76
+ ```
77
+
78
+ # Why It's benchmark score is lower than preview version?
79
+
80
+ **Apparently**, Preview model uses Alpaca Style prompt which has no pre-fix. But ChatML do.
ggml-model-q4_k_m.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59a29596cbe27a2a0e646d82cb559d63e8e08b3f6e303b00d442613028fb4110
3
+ size 4368450240