Text Generation
GGUF
English
creative
creative writing
fiction writing
plot generation
sub-plot generation
story generation
scene continue
storytelling
fiction story
science fiction
romance
all genres
story
writing
vivid prosing
vivid writing
fiction
roleplaying
bfloat16
brainstorm 40x
swearing
rp
horror
llama2
mergekit
Inference Endpoints
Update README.md
Browse files
README.md
CHANGED
@@ -80,71 +80,6 @@ Example outputs below.
|
|
80 |
- If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
|
81 |
- Source code for this model will be uploaded at a separate repo shortly.
|
82 |
|
83 |
-
<B>Brainstorm 40x</B>
|
84 |
-
|
85 |
-
The BRAINSTORM process was developed by David_AU.
|
86 |
-
|
87 |
-
Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
|
88 |
-
scientific paper : Progressive LLaMA with Block Expansion </a>.
|
89 |
-
|
90 |
-
However I went in a completely different direction from what was outlined in this paper.
|
91 |
-
|
92 |
-
I developed a process where the conclusion layer of a model is duplicated and calibrated, in the case of this model 40 times.
|
93 |
-
|
94 |
-
This is a delicate process, with umm... a lot of rules.
|
95 |
-
|
96 |
-
For this model in particular Brainstorm is mapped as blocks, with "intended disruption" to alter
|
97 |
-
and extend the power of the root model. Each layer/block interacts with each other block.
|
98 |
-
|
99 |
-
(there is more going on here too, this is rough summary)
|
100 |
-
|
101 |
-
The goal here is creative : prose uniqueness first and foremost.
|
102 |
-
|
103 |
-
Other brainstorm methods address logic/problem solving augmentation.
|
104 |
-
|
105 |
-
What is "Brainstorm" ?
|
106 |
-
|
107 |
-
The reasoning center of an LLM is taken apart, reassembled, and expanded.
|
108 |
-
|
109 |
-
In this case for this model: 40 times
|
110 |
-
|
111 |
-
Then these centers are individually calibrated. These "centers" also interact with each other.
|
112 |
-
This introduces subtle changes into the reasoning process.
|
113 |
-
The calibrations further adjust - dial up or down - these "changes" further.
|
114 |
-
The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
|
115 |
-
|
116 |
-
The core aim of this process is to increase the model's detail, concept and connection to the "world",
|
117 |
-
general concept connections, prose quality and prose length without affecting instruction following.
|
118 |
-
|
119 |
-
This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
|
120 |
-
|
121 |
-
Here are some of the enhancements this process brings to the model's performance:
|
122 |
-
|
123 |
-
- Prose generation seems more focused on the moment to moment.
|
124 |
-
- Sometimes there will be "preamble" and/or foreshadowing present.
|
125 |
-
- Fewer or no "cliches"
|
126 |
-
- Better overall prose and/or more complex / nuanced prose.
|
127 |
-
- A greater sense of nuance on all levels.
|
128 |
-
- Coherence is stronger.
|
129 |
-
- Description is more detailed, and connected closer to the content.
|
130 |
-
- Simile and Metaphors are stronger and better connected to the prose, story, and character.
|
131 |
-
- Sense of "there" / in the moment is enhanced.
|
132 |
-
- Details are more vivid, and there are more of them.
|
133 |
-
- Prose generation length can be long to extreme.
|
134 |
-
- Emotional engagement is stronger.
|
135 |
-
- The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
|
136 |
-
- The MORE instructions and/or details you provide the more strongly the model will respond.
|
137 |
-
- Depending on the model "voice" may be more "human" vs original model's "voice".
|
138 |
-
|
139 |
-
Other "lab" observations:
|
140 |
-
|
141 |
-
- This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
|
142 |
-
- However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
|
143 |
-
- From lab testing it seems to ponder, and consider more carefully roughly speaking.
|
144 |
-
- You could say this process sharpens the model's focus on it's task(s) at a deeper level.
|
145 |
-
|
146 |
-
The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
|
147 |
-
|
148 |
<B>Special Operations Notice:</B>
|
149 |
|
150 |
This is a slightly experimental model, and as a result it may "glitch" from time to time - the most common is
|
@@ -565,3 +500,70 @@ Ishiwa had done his duty. He had sent the message. He had made a difference.
|
|
565 |
|
566 |
---
|
567 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
80 |
- If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
|
81 |
- Source code for this model will be uploaded at a separate repo shortly.
|
82 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
83 |
<B>Special Operations Notice:</B>
|
84 |
|
85 |
This is a slightly experimental model, and as a result it may "glitch" from time to time - the most common is
|
|
|
500 |
|
501 |
---
|
502 |
|
503 |
+
<h2>What is Brainstorm?</h2>
|
504 |
+
|
505 |
+
<B>Brainstorm 40x</B>
|
506 |
+
|
507 |
+
The BRAINSTORM process was developed by David_AU.
|
508 |
+
|
509 |
+
Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
|
510 |
+
scientific paper : Progressive LLaMA with Block Expansion </a>.
|
511 |
+
|
512 |
+
However I went in a completely different direction from what was outlined in this paper.
|
513 |
+
|
514 |
+
I developed a process where the conclusion layer of a model is duplicated and calibrated, in the case of this model 40 times.
|
515 |
+
|
516 |
+
This is a delicate process, with umm... a lot of rules.
|
517 |
+
|
518 |
+
For this model in particular Brainstorm is mapped as blocks, with "intended disruption" to alter
|
519 |
+
and extend the power of the root model. Each layer/block interacts with each other block.
|
520 |
+
|
521 |
+
(there is more going on here too, this is rough summary)
|
522 |
+
|
523 |
+
The goal here is creative : prose uniqueness first and foremost.
|
524 |
+
|
525 |
+
Other brainstorm methods address logic/problem solving augmentation.
|
526 |
+
|
527 |
+
What is "Brainstorm" ?
|
528 |
+
|
529 |
+
The reasoning center of an LLM is taken apart, reassembled, and expanded.
|
530 |
+
|
531 |
+
In this case for this model: 40 times
|
532 |
+
|
533 |
+
Then these centers are individually calibrated. These "centers" also interact with each other.
|
534 |
+
This introduces subtle changes into the reasoning process.
|
535 |
+
The calibrations further adjust - dial up or down - these "changes" further.
|
536 |
+
The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
|
537 |
+
|
538 |
+
The core aim of this process is to increase the model's detail, concept and connection to the "world",
|
539 |
+
general concept connections, prose quality and prose length without affecting instruction following.
|
540 |
+
|
541 |
+
This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
|
542 |
+
|
543 |
+
Here are some of the enhancements this process brings to the model's performance:
|
544 |
+
|
545 |
+
- Prose generation seems more focused on the moment to moment.
|
546 |
+
- Sometimes there will be "preamble" and/or foreshadowing present.
|
547 |
+
- Fewer or no "cliches"
|
548 |
+
- Better overall prose and/or more complex / nuanced prose.
|
549 |
+
- A greater sense of nuance on all levels.
|
550 |
+
- Coherence is stronger.
|
551 |
+
- Description is more detailed, and connected closer to the content.
|
552 |
+
- Simile and Metaphors are stronger and better connected to the prose, story, and character.
|
553 |
+
- Sense of "there" / in the moment is enhanced.
|
554 |
+
- Details are more vivid, and there are more of them.
|
555 |
+
- Prose generation length can be long to extreme.
|
556 |
+
- Emotional engagement is stronger.
|
557 |
+
- The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
|
558 |
+
- The MORE instructions and/or details you provide the more strongly the model will respond.
|
559 |
+
- Depending on the model "voice" may be more "human" vs original model's "voice".
|
560 |
+
|
561 |
+
Other "lab" observations:
|
562 |
+
|
563 |
+
- This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
|
564 |
+
- However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
|
565 |
+
- From lab testing it seems to ponder, and consider more carefully roughly speaking.
|
566 |
+
- You could say this process sharpens the model's focus on it's task(s) at a deeper level.
|
567 |
+
|
568 |
+
The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
|
569 |
+
|