DavidAU commited on
Commit
cbb7e45
·
verified ·
1 Parent(s): cd22b44

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +67 -65
README.md CHANGED
@@ -80,71 +80,6 @@ Example outputs below.
80
  - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
81
  - Source code for this model will be uploaded at a separate repo shortly.
82
 
83
- <B>Brainstorm 40x</B>
84
-
85
- The BRAINSTORM process was developed by David_AU.
86
-
87
- Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
88
- scientific paper : Progressive LLaMA with Block Expansion </a>.
89
-
90
- However I went in a completely different direction from what was outlined in this paper.
91
-
92
- I developed a process where the conclusion layer of a model is duplicated and calibrated, in the case of this model 40 times.
93
-
94
- This is a delicate process, with umm... a lot of rules.
95
-
96
- For this model in particular Brainstorm is mapped as blocks, with "intended disruption" to alter
97
- and extend the power of the root model. Each layer/block interacts with each other block.
98
-
99
- (there is more going on here too, this is rough summary)
100
-
101
- The goal here is creative : prose uniqueness first and foremost.
102
-
103
- Other brainstorm methods address logic/problem solving augmentation.
104
-
105
- What is "Brainstorm" ?
106
-
107
- The reasoning center of an LLM is taken apart, reassembled, and expanded.
108
-
109
- In this case for this model: 40 times
110
-
111
- Then these centers are individually calibrated. These "centers" also interact with each other.
112
- This introduces subtle changes into the reasoning process.
113
- The calibrations further adjust - dial up or down - these "changes" further.
114
- The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
115
-
116
- The core aim of this process is to increase the model's detail, concept and connection to the "world",
117
- general concept connections, prose quality and prose length without affecting instruction following.
118
-
119
- This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
120
-
121
- Here are some of the enhancements this process brings to the model's performance:
122
-
123
- - Prose generation seems more focused on the moment to moment.
124
- - Sometimes there will be "preamble" and/or foreshadowing present.
125
- - Fewer or no "cliches"
126
- - Better overall prose and/or more complex / nuanced prose.
127
- - A greater sense of nuance on all levels.
128
- - Coherence is stronger.
129
- - Description is more detailed, and connected closer to the content.
130
- - Simile and Metaphors are stronger and better connected to the prose, story, and character.
131
- - Sense of "there" / in the moment is enhanced.
132
- - Details are more vivid, and there are more of them.
133
- - Prose generation length can be long to extreme.
134
- - Emotional engagement is stronger.
135
- - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
136
- - The MORE instructions and/or details you provide the more strongly the model will respond.
137
- - Depending on the model "voice" may be more "human" vs original model's "voice".
138
-
139
- Other "lab" observations:
140
-
141
- - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
142
- - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
143
- - From lab testing it seems to ponder, and consider more carefully roughly speaking.
144
- - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
145
-
146
- The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
147
-
148
  <B>Special Operations Notice:</B>
149
 
150
  This is a slightly experimental model, and as a result it may "glitch" from time to time - the most common is
@@ -565,3 +500,70 @@ Ishiwa had done his duty. He had sent the message. He had made a difference.
565
 
566
  ---
567
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
80
  - If you use rope to extend context, increase temp AND instructions detail levels to compensate for "rope issues".
81
  - Source code for this model will be uploaded at a separate repo shortly.
82
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
83
  <B>Special Operations Notice:</B>
84
 
85
  This is a slightly experimental model, and as a result it may "glitch" from time to time - the most common is
 
500
 
501
  ---
502
 
503
+ <h2>What is Brainstorm?</h2>
504
+
505
+ <B>Brainstorm 40x</B>
506
+
507
+ The BRAINSTORM process was developed by David_AU.
508
+
509
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
510
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
511
+
512
+ However I went in a completely different direction from what was outlined in this paper.
513
+
514
+ I developed a process where the conclusion layer of a model is duplicated and calibrated, in the case of this model 40 times.
515
+
516
+ This is a delicate process, with umm... a lot of rules.
517
+
518
+ For this model in particular Brainstorm is mapped as blocks, with "intended disruption" to alter
519
+ and extend the power of the root model. Each layer/block interacts with each other block.
520
+
521
+ (there is more going on here too, this is rough summary)
522
+
523
+ The goal here is creative : prose uniqueness first and foremost.
524
+
525
+ Other brainstorm methods address logic/problem solving augmentation.
526
+
527
+ What is "Brainstorm" ?
528
+
529
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
530
+
531
+ In this case for this model: 40 times
532
+
533
+ Then these centers are individually calibrated. These "centers" also interact with each other.
534
+ This introduces subtle changes into the reasoning process.
535
+ The calibrations further adjust - dial up or down - these "changes" further.
536
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
537
+
538
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
539
+ general concept connections, prose quality and prose length without affecting instruction following.
540
+
541
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
542
+
543
+ Here are some of the enhancements this process brings to the model's performance:
544
+
545
+ - Prose generation seems more focused on the moment to moment.
546
+ - Sometimes there will be "preamble" and/or foreshadowing present.
547
+ - Fewer or no "cliches"
548
+ - Better overall prose and/or more complex / nuanced prose.
549
+ - A greater sense of nuance on all levels.
550
+ - Coherence is stronger.
551
+ - Description is more detailed, and connected closer to the content.
552
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
553
+ - Sense of "there" / in the moment is enhanced.
554
+ - Details are more vivid, and there are more of them.
555
+ - Prose generation length can be long to extreme.
556
+ - Emotional engagement is stronger.
557
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
558
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
559
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
560
+
561
+ Other "lab" observations:
562
+
563
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
564
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
565
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
566
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
567
+
568
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
569
+