Triangle104
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -21,6 +21,97 @@ tags:
|
|
21 |
This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
22 |
Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
|
23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
## Use with llama.cpp
|
25 |
Install llama.cpp through brew (works on Mac and Linux)
|
26 |
|
|
|
21 |
This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
|
22 |
Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
|
23 |
|
24 |
+
---
|
25 |
+
Model details:
|
26 |
+
-
|
27 |
+
|
28 |
+
A RP/storywriting specialist model, full-parameter finetune of Qwen2.5-7B on mixture of synthetic and natural data.
|
29 |
+
|
30 |
+
It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve
|
31 |
+
versatility, creativity and "flavor" of the resulting model.
|
32 |
+
|
33 |
+
|
34 |
+
|
35 |
+
|
36 |
+
|
37 |
+
|
38 |
+
Version 0.1 notes:
|
39 |
+
Dataset was deduped and cleaned from
|
40 |
+
version 0.0, and learning rate was adjusted. Resulting model seems to be
|
41 |
+
stabler, and 0.0 problems with handling short inputs and min_p sampling
|
42 |
+
seem to be mostly gone.
|
43 |
+
|
44 |
+
Will be retrained once more, because this run crashed around e1.2 (out
|
45 |
+
of 3) (thanks, DeepSpeed, really appreciate it), and it's still
|
46 |
+
somewhat undertrained as a result.
|
47 |
+
|
48 |
+
|
49 |
+
|
50 |
+
|
51 |
+
|
52 |
+
|
53 |
+
|
54 |
+
Prompt format is ChatML.
|
55 |
+
|
56 |
+
|
57 |
+
|
58 |
+
Recommended sampler values:
|
59 |
+
|
60 |
+
|
61 |
+
Temperature: 0.87
|
62 |
+
Top-P: 0.81
|
63 |
+
Repetition Penalty: 1.03
|
64 |
+
|
65 |
+
|
66 |
+
Model appears to prefer lower temperatures (at least 0.9 and lower). Min-P seems to work now, as well.
|
67 |
+
|
68 |
+
|
69 |
+
|
70 |
+
Recommended SillyTavern presets (via CalamitousFelicitousness):
|
71 |
+
|
72 |
+
|
73 |
+
|
74 |
+
Context
|
75 |
+
Instruct and System Prompt
|
76 |
+
|
77 |
+
|
78 |
+
|
79 |
+
|
80 |
+
|
81 |
+
|
82 |
+
|
83 |
+
|
84 |
+
Training data:
|
85 |
+
|
86 |
+
|
87 |
+
|
88 |
+
Celeste 70B 0.1 data mixture minus Opus Instruct subset. See that model's card for details.
|
89 |
+
Kalomaze's Opus_Instruct_25k dataset, filtered for refusals.
|
90 |
+
A subset (1k rows) of ChatGPT-4o-WritingPrompts by Gryphe
|
91 |
+
A subset (2k rows) of Sonnet3.5-Charcards-Roleplay by Gryphe
|
92 |
+
A cleaned subset (~3k rows) of shortstories_synthlabels by Auri
|
93 |
+
Synthstruct and SynthRP datasets by Epiculous
|
94 |
+
|
95 |
+
|
96 |
+
|
97 |
+
Training time and hardware:
|
98 |
+
|
99 |
+
|
100 |
+
|
101 |
+
2 days on 4x3090Ti (locally)
|
102 |
+
|
103 |
+
|
104 |
+
|
105 |
+
|
106 |
+
|
107 |
+
Model was trained by Kearm and Auri.
|
108 |
+
|
109 |
+
Special thanks:
|
110 |
+
to Gryphe, Lemmy, Kalomaze, Nopm and Epiculous for the data
|
111 |
+
to Alpindale for helping with FFT config for Qwen2.5
|
112 |
+
and to InfermaticAI's community for their continued support for our endeavors
|
113 |
+
|
114 |
+
---
|
115 |
## Use with llama.cpp
|
116 |
Install llama.cpp through brew (works on Mac and Linux)
|
117 |
|