DavidAU commited on
Commit
cb4f4c1
·
verified ·
1 Parent(s): 12fa61c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +61 -1
README.md CHANGED
@@ -47,7 +47,7 @@ pipeline_tag: text-generation
47
  <img src="9b-llama.jpg" style="float:right; width:300px; height:300px; padding:10px;">
48
 
49
  First in the new Llama 3.2 models at 9B parameters and built with "
50
- Hastagaras/L3.2-JametMini-3B-MK.III", "chuanli11/Llama-3.2-3B-Instruct-uncensored" and the Brainstorm 40x adapter by DavidAU.
51
 
52
  This is an 81 layer [733 tensors], Brainstorm 40x enhanced (changes token generation, enhances performance), uncensored monster.
53
 
@@ -456,3 +456,63 @@ But she did reach out and took his hand – and they went on that walk together.
456
 
457
 
458
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  <img src="9b-llama.jpg" style="float:right; width:300px; height:300px; padding:10px;">
48
 
49
  First in the new Llama 3.2 models at 9B parameters and built with "
50
+ Hastagaras/L3.2-JametMini-3B-MK.III", "chuanli11/Llama-3.2-3B-Instruct-uncensored" and the Brainstorm 40x adapter (notes at bottom of the page) by DavidAU.
51
 
52
  This is an 81 layer [733 tensors], Brainstorm 40x enhanced (changes token generation, enhances performance), uncensored monster.
53
 
 
456
 
457
 
458
 
459
+ ---
460
+
461
+ <H2>What is Brainstorm?</H2>
462
+
463
+ ---
464
+
465
+ <B>Brainstorm 40x</B>
466
+
467
+ The BRAINSTORM process was developed by David_AU.
468
+
469
+ Some of the core principals behind this process are discussed in this <a href="https://arxiv.org/pdf/2401.02415">
470
+ scientific paper : Progressive LLaMA with Block Expansion </a>.
471
+
472
+ However I went in a completely different direction from what was outlined in this paper.
473
+
474
+ What is "Brainstorm" ?
475
+
476
+ The reasoning center of an LLM is taken apart, reassembled, and expanded.
477
+
478
+ In this case for this model: 40 times
479
+
480
+ Then these centers are individually calibrated. These "centers" also interact with each other.
481
+ This introduces subtle changes into the reasoning process.
482
+ The calibrations further adjust - dial up or down - these "changes" further.
483
+ The number of centers (5x,10x etc) allow more "tuning points" to further customize how the model reasons so to speak.
484
+
485
+ The core aim of this process is to increase the model's detail, concept and connection to the "world",
486
+ general concept connections, prose quality and prose length without affecting instruction following.
487
+
488
+ This will also enhance any creative use case(s) of any kind, including "brainstorming", creative art form(s) and like case uses.
489
+
490
+ Here are some of the enhancements this process brings to the model's performance:
491
+
492
+ - Prose generation seems more focused on the moment to moment.
493
+ - Sometimes there will be "preamble" and/or foreshadowing present.
494
+ - Fewer or no "cliches"
495
+ - Better overall prose and/or more complex / nuanced prose.
496
+ - A greater sense of nuance on all levels.
497
+ - Coherence is stronger.
498
+ - Description is more detailed, and connected closer to the content.
499
+ - Simile and Metaphors are stronger and better connected to the prose, story, and character.
500
+ - Sense of "there" / in the moment is enhanced.
501
+ - Details are more vivid, and there are more of them.
502
+ - Prose generation length can be long to extreme.
503
+ - Emotional engagement is stronger.
504
+ - The model will take FEWER liberties vs a normal model: It will follow directives more closely but will "guess" less.
505
+ - The MORE instructions and/or details you provide the more strongly the model will respond.
506
+ - Depending on the model "voice" may be more "human" vs original model's "voice".
507
+
508
+ Other "lab" observations:
509
+
510
+ - This process does not, in my opinion, make the model 5x or 10x "smarter" - if only that was true!
511
+ - However, a change in "IQ" was not an issue / a priority, and was not tested or calibrated for so to speak.
512
+ - From lab testing it seems to ponder, and consider more carefully roughly speaking.
513
+ - You could say this process sharpens the model's focus on it's task(s) at a deeper level.
514
+
515
+ The process to modify the model occurs at the root level - source files level. The model can quanted as a GGUF, EXL2, AWQ etc etc.
516
+
517
+ ---
518
+