gagan3012 commited on
Commit
015dae6
1 Parent(s): 9ff76e9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +37 -37
README.md CHANGED
@@ -5,47 +5,47 @@ tags:
5
  - mixtral
6
  ---
7
 
8
- # MetaModel_moe
9
 
10
- This model is a Mixure of Experts (MoE) made with [mergekit](https://github.com/cg123/mergekit) (mixtral branch). It uses the following base models:
11
- * [gagan3012/MetaModel](https://huggingface.co/gagan3012/MetaModel)
12
- * [jeonsworld/CarbonVillain-en-10.7B-v2](https://huggingface.co/jeonsworld/CarbonVillain-en-10.7B-v2)
13
- * [jeonsworld/CarbonVillain-en-10.7B-v4](https://huggingface.co/jeonsworld/CarbonVillain-en-10.7B-v4)
14
- * [TomGrc/FusionNet_linear](https://huggingface.co/TomGrc/FusionNet_linear)
15
 
16
- ## 🧩 Configuration
17
 
18
- ```yaml
19
  base_model: gagan3012/MetaModel
20
  gate_mode: hidden
21
  dtype: bfloat16
22
  experts:
23
- - source_model: gagan3012/MetaModel
24
- - source_model: jeonsworld/CarbonVillain-en-10.7B-v2
25
- - source_model: jeonsworld/CarbonVillain-en-10.7B-v4
26
- - source_model: TomGrc/FusionNet_linear
27
- ```
28
-
29
- ## 💻 Usage
30
-
31
- ```python
32
- !pip install -qU transformers bitsandbytes accelerate
33
-
34
- from transformers import AutoTokenizer
35
- import transformers
36
- import torch
37
-
38
- model = "gagan3012/MetaModel_moe"
39
-
40
- tokenizer = AutoTokenizer.from_pretrained(model)
41
- pipeline = transformers.pipeline(
42
- "text-generation",
43
- model=model,
44
- model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
45
- )
46
-
47
- messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
48
- prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
49
- outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
50
- print(outputs[0]["generated_text"])
51
- ```
 
5
  - mixtral
6
  ---
7
 
8
+ # MetaModel_moe
9
 
10
+ This model is a Mixure of Experts (MoE) made with [mergekit](https://github.com/cg123/mergekit) (mixtral branch). It uses the following base models:
11
+ * [gagan3012/MetaModel](https://huggingface.co/gagan3012/MetaModel)
12
+ * [jeonsworld/CarbonVillain-en-10.7B-v2](https://huggingface.co/jeonsworld/CarbonVillain-en-10.7B-v2)
13
+ * [jeonsworld/CarbonVillain-en-10.7B-v4](https://huggingface.co/jeonsworld/CarbonVillain-en-10.7B-v4)
14
+ * [TomGrc/FusionNet_linear](https://huggingface.co/TomGrc/FusionNet_linear)
15
 
16
+ ## 🧩 Configuration
17
 
18
+ ```yaml
19
  base_model: gagan3012/MetaModel
20
  gate_mode: hidden
21
  dtype: bfloat16
22
  experts:
23
+ - source_model: gagan3012/MetaModel
24
+ - source_model: jeonsworld/CarbonVillain-en-10.7B-v2
25
+ - source_model: jeonsworld/CarbonVillain-en-10.7B-v4
26
+ - source_model: TomGrc/FusionNet_linear
27
+ ```
28
+
29
+ ## 💻 Usage
30
+
31
+ ```python
32
+ !pip install -qU transformers bitsandbytes accelerate
33
+
34
+ from transformers import AutoTokenizer
35
+ import transformers
36
+ import torch
37
+
38
+ model = "gagan3012/MetaModel_moe"
39
+
40
+ tokenizer = AutoTokenizer.from_pretrained(model)
41
+ pipeline = transformers.pipeline(
42
+ "text-generation",
43
+ model=model,
44
+ model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True},
45
+ )
46
+
47
+ messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}]
48
+ prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
49
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
50
+ print(outputs[0]["generated_text"])
51
+ ```