DavidAU commited on
Commit
96afb32
1 Parent(s): ec5e60b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -0
README.md CHANGED
@@ -138,6 +138,33 @@ A regen will usually correct any issues.
138
  Some of the "censorship" of the original Llama 3.2 3B Instruct model is still present in this model.
139
  For some generations (ie to get it to "swear") you may need to regen it 1-2+ times to get the model to "obey".
140
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
141
  <B>Model Template:</B>
142
 
143
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 128k / 131072.
 
138
  Some of the "censorship" of the original Llama 3.2 3B Instruct model is still present in this model.
139
  For some generations (ie to get it to "swear") you may need to regen it 1-2+ times to get the model to "obey".
140
 
141
+ <B>Settings: CHAT / ROLEPLAY and/or SMOOTHER operation of this model:</B>
142
+
143
+ In "KoboldCpp" or "oobabooga/text-generation-webui" or "Silly Tavern" ;
144
+
145
+ Set the "Smoothing_factor" to 1.5 to 2.5
146
+
147
+ : in KoboldCpp -> Settings->Samplers->Advanced-> "Smooth_F"
148
+
149
+ : in text-generation-webui -> parameters -> lower right.
150
+
151
+ : In Silly Tavern this is called: "Smoothing"
152
+
153
+
154
+ NOTE: For "text-generation-webui"
155
+
156
+ -> if using GGUFs you need to use "llama_HF" (which involves downloading some config files from the SOURCE version of this model)
157
+
158
+ Source versions (and config files) of my models are here:
159
+
160
+ https://huggingface.co/collections/DavidAU/d-au-source-files-for-gguf-exl2-awq-gptq-hqq-etc-etc-66b55cb8ba25f914cbf210be
161
+
162
+ OTHER OPTIONS:
163
+
164
+ - Increase rep pen to 1.1 to 1.15 (you don't need to do this if you use "smoothing_factor")
165
+
166
+ - If the interface/program you are using to run AI MODELS supports "Quadratic Sampling" ("smoothing") just make the adjustment as noted.
167
+
168
  <B>Model Template:</B>
169
 
170
  This is a LLAMA3 model, and requires Llama3 template, but may work with other template(s) and has maximum context of 128k / 131072.