leafspark commited on
Commit
3a58e69
·
verified ·
1 Parent(s): c41590a

docs: add model card

Browse files
Files changed (1) hide show
  1. README.md +63 -5
README.md CHANGED
@@ -1,5 +1,63 @@
1
- ---
2
- license: other
3
- license_name: tongyi-qianwen
4
- license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE
5
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: other
3
+ license_name: tongyi-qianwen
4
+ license_link: https://huggingface.co/Qwen/Qwen2-72B-Instruct/blob/main/LICENSE
5
+ pipeline_tag: text-generation
6
+ language:
7
+ - en
8
+ - zh
9
+ library_name: transformers
10
+ tags:
11
+ - mergekit
12
+ - llama
13
+ ---
14
+
15
+ # FeatherLlama-72B-v0.1
16
+
17
+ ## Model Description
18
+ FeatherLlama is a 72B parameter language model created through a merge of Qwen2-72B-Instruct, calme2.1-72b, and magnum-72b-v1 using `model_stock`.
19
+
20
+ This is converted from [leafspark/FeatherQwen2-72B-v0.1](https://huggingface.co/leafspark/FeatherQwen2-72B-v0.1)
21
+
22
+ ## Features
23
+ - 72 billion parameters
24
+ - Sharded in 31 files (unlike FeatherQwen2, which has 1,043 shards due to the merging process)
25
+ - Combines Magnum prose with Calam smarts
26
+ - Llamaified for easy use
27
+
28
+ ## Technical Specifications
29
+
30
+ ### Architecture
31
+ - `LlamaForCasualLM`
32
+ - Models: Qwen2-72B-Instruct (base), calme2.1-72b, magnum-72b-v1
33
+ - Merged layers: 80
34
+ - Total tensors: 1,043
35
+
36
+ ### Tensor Distribution
37
+ - Attention layers: 560 files
38
+ - MLP layers: 240 files
39
+ - Layer norms: 160 files
40
+ - Miscellaneous (embeddings, output): 83 files
41
+
42
+ ### Merging
43
+ Custom script utilizing safetensors library.
44
+
45
+ ## Usage
46
+
47
+ ### Loading the Model
48
+ ```python
49
+ from transformers import AutoModelForCausalLM, AutoTokenizer
50
+ import torch
51
+
52
+ model = AutoModelForCausalLM.from_pretrained("leafspark/FeatherLlama-72B-v0.1",
53
+ device_map="auto",
54
+ torch_dtype=torch.float16)
55
+ tokenizer = AutoTokenizer.from_pretrained("leafspark/FeatherLlama-72B-v0.1")
56
+ ```
57
+ ### GGUFs
58
+
59
+ Find them here: [leafspark/FeatherLlama-72B-v0.1-GGUF](https://huggingface.co/leafspark/FeatherLlama-72B-v0.1-GGUF)
60
+
61
+ ### Hardware Requirements
62
+ - Minimum ~140GB of storage
63
+ - ~140GB VRAM