Update README.md
Browse files
README.md
CHANGED
@@ -105,7 +105,14 @@ This is `BEE-spoke-data/smol_llama-220M-GQA` fine-tuned for code generation on:
|
|
105 |
|
106 |
This model (and the base model) were both trained using ctx length 2048.
|
107 |
|
108 |
-
|
109 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
110 |
|
111 |
---
|
|
|
105 |
|
106 |
This model (and the base model) were both trained using ctx length 2048.
|
107 |
|
108 |
+
## examples
|
109 |
|
110 |
+
> Example script for inference testing: [here](https://gist.github.com/pszemraj/c7738f664a64b935a558974d23a7aa8c)
|
111 |
+
|
112 |
+
It has its limitations at 220M, but seems decent for single-line or docstring generation, and/or being used for speculative decoding for such purposes.
|
113 |
+
|
114 |
+
|
115 |
+
|
116 |
+

|
117 |
|
118 |
---
|