parameters guide
samplers guide
model generation
role play settings
quant selection
arm quants
iq quants vs q quants
optimal model setting
gibberish fixes
coherence
instructing following
quality generation
chat settings
quality settings
llamacpp server
llamacpp
lmstudio
sillytavern
koboldcpp
backyard
ollama
model generation steering
steering
model generation fixes
text generation webui
ggufs
exl2
full precision
quants
imatrix
neo imatrix
Update README.md
Browse files
README.md
CHANGED
@@ -580,19 +580,37 @@ However, you should also check / test operation of:
|
|
580 |
a] Affects per token generation:
|
581 |
|
582 |
- top_a
|
583 |
-
- epsilon_cutoff
|
584 |
-
- eta_cutoff
|
585 |
-
- no_repeat_ngram_size
|
586 |
|
587 |
b] Affects generation including phrase, sentence, paragraph and entire generation:
|
588 |
|
589 |
-
- no_repeat_ngram_size
|
590 |
-
- encoder_repetition_penalty
|
591 |
-
- guidance_scale (with "Negative prompt" ) => this is like a pre-prompt/system role prompt.
|
592 |
- Disabling (BOS TOKEN) this can make the replies more creative.
|
593 |
- Custom stopping strings
|
594 |
|
595 |
-
Note:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
596 |
|
597 |
|
598 |
<B>MAIN ADVANCED SAMPLERS (affects per token AND overall generation): </B>
|
|
|
580 |
a] Affects per token generation:
|
581 |
|
582 |
- top_a
|
583 |
+
- epsilon_cutoff - see note 4
|
584 |
+
- eta_cutoff - see note 4
|
585 |
+
- no_repeat_ngram_size - see note 1.
|
586 |
|
587 |
b] Affects generation including phrase, sentence, paragraph and entire generation:
|
588 |
|
589 |
+
- no_repeat_ngram_size - see note 1.
|
590 |
+
- encoder_repetition_penalty "Hallucinations filter" - see note #2.
|
591 |
+
- guidance_scale (with "Negative prompt" ) => this is like a pre-prompt/system role prompt - see note #3.
|
592 |
- Disabling (BOS TOKEN) this can make the replies more creative.
|
593 |
- Custom stopping strings
|
594 |
|
595 |
+
Note 1:
|
596 |
+
|
597 |
+
"no_repeat_ngram_size" appears in both because it can impact per token OR per phrase depending on settings. This can also drastically affect sentence,
|
598 |
+
paragraph and general flow of the output.
|
599 |
+
|
600 |
+
Note 2:
|
601 |
+
|
602 |
+
This parameter if set to LESS than 1 causing the model to "jump" around a lot more , whereas above 1 causes the model to focus more on the immediate surroundings.
|
603 |
+
|
604 |
+
If the model is crafting a "scene", a setting of less than 1 causes the model to jump around the room, outside, etc etc ; if less than 1 then it focuses the model more on
|
605 |
+
the moment, the immediate surroundings, the POV character and details in the setting.
|
606 |
+
|
607 |
+
Note 3:
|
608 |
+
|
609 |
+
This is a powerful method to send instructions / directives to the model on how to process your prompt(s) each time. See [ https://arxiv.org/pdf/2306.17806 ]
|
610 |
+
|
611 |
+
Note 4:
|
612 |
+
|
613 |
+
These control selection of tokens, in some case providing more relevant and/or more options. See [ https://arxiv.org/pdf/2210.15191 ]
|
614 |
|
615 |
|
616 |
<B>MAIN ADVANCED SAMPLERS (affects per token AND overall generation): </B>
|