DavidAU commited on
Commit
990485c
1 Parent(s): 2257b15

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md CHANGED
@@ -136,6 +136,33 @@ Short prompts with some rep pen/temp combinations may lead to longer than expect
136
 
137
  A regen will usually correct any issues.
138
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
139
  <B>Model Template:</B>
140
 
141
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 128k / 131072.
 
136
 
137
  A regen will usually correct any issues.
138
 
139
+ <B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
140
+
141
+ In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
142
+
143
+ Set the "Smoothing_factor" to 1.5 to 2.5
144
+
145
+ : in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
146
+
147
+ : in text-generation-webui -> parameters -> lower right.
148
+
149
+ : In Silly Tavern this is called: "Smoothing"
150
+
151
+
152
+ NOTE: For "text-generation-webui"
153
+
154
+ -> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
155
+
156
+ Source versions (and config files) of my models are here:
157
+
158
+ https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
159
+
160
+ OTHER OPTIONS:
161
+
162
+ - Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
163
+
164
+ - If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
165
+
166
  <B>Model Template:</B>
167
 
168
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 128k / 131072.