LeroyDyer commited on
Commit
7ae8188
1 Parent(s): 29d68a7

Delete README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -54
README.md DELETED
@@ -1,54 +0,0 @@
1
- ---
2
- base_model:
3
- - LeroyDyer/Mixtral_AI_CyberBrain_3_0
4
- - ezelikman/quietstar-8-ahead
5
- library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
-
10
- ---
11
- # merge
12
-
13
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
14
-
15
- ## Merge Details
16
- ### Merge Method
17
-
18
- This model was merged using the SLERP merge method.
19
-
20
- ### Models Merged
21
-
22
- The following models were included in the merge:
23
- * [LeroyDyer/Mixtral_AI_CyberBrain_3_0](https://huggingface.co/LeroyDyer/Mixtral_AI_CyberBrain_3_0)
24
- * [ezelikman/quietstar-8-ahead](https://huggingface.co/ezelikman/quietstar-8-ahead)
25
-
26
- ### Configuration
27
-
28
- The following YAML configuration was used to produce this model:
29
-
30
- ```yaml
31
-
32
- slices:
33
- - sources:
34
- - model: LeroyDyer/Mixtral_AI_CyberBrain_3_0
35
- layer_range: [0, 32]
36
- - model: ezelikman/quietstar-8-ahead
37
- layer_range: [0, 32]
38
- # or, the equivalent models: syntax:
39
- # models:
40
- # - model: mistralai/Mistral-7B-Instruct-v0.2
41
- # LaRGER MODEL MUST BE BASE
42
- # - model: yanismiraoui/Yarn-Mistral-7b-128k-sharded
43
- merge_method: slerp
44
- base_model: ezelikman/quietstar-8-ahead
45
- parameters:
46
- t:
47
- - filter: self_attn
48
- value: [0.3, 0.6, 0.3786, 0.6, 0.6]
49
- - filter: mlp
50
- value: [0.7, 0.4, 0.6, 0.4, 0.7]
51
- - value: 0.5 # fallback for rest of tensors
52
- dtype: float16
53
-
54
- ```