jaspercatapang
commited on
Commit
·
f08d6f6
1
Parent(s):
00bc4a5
Update README.md
Browse files
README.md
CHANGED
@@ -1,16 +1,34 @@
|
|
1 |
---
|
2 |
pipeline_tag: text-generation
|
|
|
3 |
---
|
4 |
|
5 |
# GodziLLa-30B 🦖
|
6 |
Released July 9, 2023
|
7 |
|
8 |
## Model Description
|
9 |
-
GodziLLa-30B is an experimental combination of various LoRAs with CalderaAI's [Lazarus-30B](https://huggingface.co/CalderaAI/30B-Lazarus). This composite model is not meant for
|
10 |
|
11 |
## Recommended Prompt Format
|
12 |
Alpaca's instruction is the recommended prompt format, but Vicuna's instruction format may also work.
|
13 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
14 |
## Ethical Considerations
|
15 |
When using GodziLLa-30B, it is important to consider the following ethical considerations:
|
16 |
|
|
|
1 |
---
|
2 |
pipeline_tag: text-generation
|
3 |
+
license: cc-by-nc-4.0
|
4 |
---
|
5 |
|
6 |
# GodziLLa-30B 🦖
|
7 |
Released July 9, 2023
|
8 |
|
9 |
## Model Description
|
10 |
+
GodziLLa-30B is an experimental combination of various proprietary Maya LoRAs with CalderaAI's [Lazarus-30B](https://huggingface.co/CalderaAI/30B-Lazarus). This composite model is not meant for any other use outside of research on competing LoRA adapter behavior. More specifically, since this is inherently a LlaMA model, **commercial use is prohibited**. This model's primary purpose is to stress test the limitations of composite LLMs and observe its performance with respect to other LLMs available on the [Open LLM Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
|
11 |
|
12 |
## Recommended Prompt Format
|
13 |
Alpaca's instruction is the recommended prompt format, but Vicuna's instruction format may also work.
|
14 |
|
15 |
+
## Usage
|
16 |
+
To use GodziLLa-30B, you are required to provide attribution in accordance with the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license. Please include the following attribution notice when utilizing GodziLLa-30B in your work:
|
17 |
+
|
18 |
+
```python
|
19 |
+
# This code uses GodziLLa-30B, a language model developed by Maya Philippines.
|
20 |
+
# The model is licensed under the Creative Commons Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) license.
|
21 |
+
# For more information, visit: https://creativecommons.org/licenses/by-nc/4.0/
|
22 |
+
|
23 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
24 |
+
|
25 |
+
tokenizer = AutoTokenizer.from_pretrained("MayaPH/GodziLLa-30B")
|
26 |
+
|
27 |
+
model = AutoModelForCausalLM.from_pretrained("MayaPH/GodziLLa-30B")
|
28 |
+
```
|
29 |
+
|
30 |
+
Please ensure that you include the relevant attribution notice in your code or any other form of usage and restrict your usage to non-commercial use to comply with the license terms.
|
31 |
+
|
32 |
## Ethical Considerations
|
33 |
When using GodziLLa-30B, it is important to consider the following ethical considerations:
|
34 |
|