Update README.md
Browse files
README.md
CHANGED
@@ -81,6 +81,8 @@ model-index:
|
|
81 |
---
|
82 |
# `stable-code-3b`
|
83 |
|
|
|
|
|
84 |
## Model Description
|
85 |
|
86 |
`stable-code-3b` is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. `stable-code-3b` is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using [BigCode's Evaluation Harness](https://github.com/bigcode-project/bigcode-evaluation-harness/tree/main).
|
@@ -184,7 +186,8 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True))
|
|
184 |
* **Model type**: `stable-code-3b` models are auto-regressive language models based on the transformer decoder architecture.
|
185 |
* **Language(s)**: English, Code
|
186 |
* **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
|
187 |
-
* **License**:
|
|
|
188 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
189 |
|
190 |
### Model Architecture
|
@@ -238,7 +241,7 @@ The model is pre-trained on the aforementioned datasets in `bfloat16` precision,
|
|
238 |
|
239 |
### Intended Use
|
240 |
|
241 |
-
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications.
|
242 |
|
243 |
### Limitations and Bias
|
244 |
|
|
|
81 |
---
|
82 |
# `stable-code-3b`
|
83 |
|
84 |
+
Please note: For commercial use, please refer to https://stability.ai/membership.
|
85 |
+
|
86 |
## Model Description
|
87 |
|
88 |
`stable-code-3b` is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. `stable-code-3b` is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using [BigCode's Evaluation Harness](https://github.com/bigcode-project/bigcode-evaluation-harness/tree/main).
|
|
|
186 |
* **Model type**: `stable-code-3b` models are auto-regressive language models based on the transformer decoder architecture.
|
187 |
* **Language(s)**: English, Code
|
188 |
* **Library**: [GPT-NeoX](https://github.com/EleutherAI/gpt-neox)
|
189 |
+
* **License**: Stability AI Non-Commercial Research Community License.
|
190 |
+
* **Commercial License**: to use this model commercially, please refer to https://stability.ai/membership
|
191 |
* **Contact**: For questions and comments about the model, please email `[email protected]`
|
192 |
|
193 |
### Model Architecture
|
|
|
241 |
|
242 |
### Intended Use
|
243 |
|
244 |
+
The model is intended to be used as a foundational base model for application-specific fine-tuning. Developers must evaluate and fine-tune the model for safe performance in downstream applications. For commercial use, please refer to https://stability.ai/membership.
|
245 |
|
246 |
### Limitations and Bias
|
247 |
|