Triangle104 commited on
Commit
fc28611
·
verified ·
1 Parent(s): b926341

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +91 -0
README.md CHANGED
@@ -21,6 +21,97 @@ tags:
21
  This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  ## Use with llama.cpp
25
  Install llama.cpp through brew (works on Mac and Linux)
26
 
 
21
  This model was converted to GGUF format from [`EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1`](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
22
  Refer to the [original model card](https://huggingface.co/EVA-UNIT-01/EVA-Qwen2.5-7B-v0.1) for more details on the model.
23
 
24
+ ---
25
+ Model details:
26
+ -
27
+
28
+ A RP/storywriting specialist model, full-parameter finetune of Qwen2.5-7B on mixture of synthetic and natural data.
29
+
30
+ It uses Celeste 70B 0.1 data mixture, greatly expanding it to improve
31
+ versatility, creativity and "flavor" of the resulting model.
32
+
33
+
34
+
35
+
36
+
37
+
38
+ Version 0.1 notes:
39
+ Dataset was deduped and cleaned from
40
+ version 0.0, and learning rate was adjusted. Resulting model seems to be
41
+ stabler, and 0.0 problems with handling short inputs and min_p sampling
42
+ seem to be mostly gone.
43
+
44
+ Will be retrained once more, because this run crashed around e1.2 (out
45
+ of 3) (thanks, DeepSpeed, really appreciate it), and it's still
46
+ somewhat undertrained as a result.
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+ Prompt format is ChatML.
55
+
56
+
57
+
58
+ Recommended sampler values:
59
+
60
+
61
+ Temperature: 0.87
62
+ Top-P: 0.81
63
+ Repetition Penalty: 1.03
64
+
65
+
66
+ Model appears to prefer lower temperatures (at least 0.9 and lower). Min-P seems to work now, as well.
67
+
68
+
69
+
70
+ Recommended SillyTavern presets (via CalamitousFelicitousness):
71
+
72
+
73
+
74
+ Context
75
+ Instruct and System Prompt
76
+
77
+
78
+
79
+
80
+
81
+
82
+
83
+
84
+ Training data:
85
+
86
+
87
+
88
+ Celeste 70B 0.1 data mixture minus Opus Instruct subset. See that model's card for details.
89
+ Kalomaze's Opus_Instruct_25k dataset, filtered for refusals.
90
+ A subset (1k rows) of ChatGPT-4o-WritingPrompts by Gryphe
91
+ A subset (2k rows) of Sonnet3.5-Charcards-Roleplay by Gryphe
92
+ A cleaned subset (~3k rows) of shortstories_synthlabels by Auri
93
+ Synthstruct and SynthRP datasets by Epiculous
94
+
95
+
96
+
97
+ Training time and hardware:
98
+
99
+
100
+
101
+ 2 days on 4x3090Ti (locally)
102
+
103
+
104
+
105
+
106
+
107
+ Model was trained by Kearm and Auri.
108
+
109
+ Special thanks:
110
+ to Gryphe, Lemmy, Kalomaze, Nopm and Epiculous for the data
111
+ to Alpindale for helping with FFT config for Qwen2.5
112
+ and to InfermaticAI's community for their continued support for our endeavors
113
+
114
+ ---
115
  ## Use with llama.cpp
116
  Install llama.cpp through brew (works on Mac and Linux)
117