LLM coping mechanisms - Part 5
pinned
145
#12 opened 7 months ago
by
Lewdiculous
Is it normal that Gemma models do not work with kv-cache?
#19 opened 5 months ago
by
SolidSnacke
[llama.cpp PR#7527] GGUF Quantized KV Support
22
#15 opened 7 months ago
by
Lewdiculous
[llama.cpp PR#6844] Custom Quantizations
6
#8 opened 8 months ago
by
Virt-io
Sampling Resources and Conjecture
44
#2 opened 8 months ago
by
Clevyby