qwp4w3hyb commited on
Commit
a6daa39
·
verified ·
1 Parent(s): 6955e8a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +45 -5
README.md CHANGED
@@ -36,14 +36,23 @@ The newest quants uploaded here need at least commit f4ab2a41476600a98067a9474ea
36
 
37
  # Original Model Card
38
 
 
 
39
  # Llama-3 8B Gradient Instruct 1048k
 
 
 
40
  Gradient incorporates your data to deploy autonomous assistants that power critical operations across your business. If you're looking to build custom AI models or agents, email us a message [email protected].
41
 
42
  For more info see our [End-to-end development service for custom LLMs and AI systems](https://gradient.ai/development-lab)
43
 
 
 
44
  This model extends LLama-3 8B's context length from 8k to > 1040K, developed by Gradient, sponsored by compute from [Crusoe Energy](https://huggingface.co/crusoeai). It demonstrates that SOTA LLMs can learn to operate on long context with minimal training by appropriately adjusting RoPE theta. We trained on 830M tokens for this stage, and 1.4B tokens total for all stages, which is < 0.01% of Llama-3's original pre-training data.
45
 
46
- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585dc9be92bc5f258156bd6/6MKLoX2ruLIaREiyb6coO.png)
 
 
47
 
48
  **Approach:**
49
 
@@ -59,7 +68,7 @@ Notably, we layered parallelism on top of Ring Attention with a custom network t
59
 
60
  **Data:**
61
 
62
- For training data, we generate long contexts by augmenting [SlimPajama](https://huggingface.co/datasets/cerebras/SlimPajama-627B).
63
 
64
  **Progressive Training Details:**
65
 
@@ -68,7 +77,7 @@ For training data, we generate long contexts by augmenting [SlimPajama](https://
68
  | Initialize From | LLaMA-3 8B| 65K | 262K | 524k |
69
  | Sequence Length 2^N | 16 | 18 | 19 | 20 |
70
  | RoPE theta | 15.3 M | 207.1 M | 1.06B | 2.80B |
71
- | Batch Size | 1 | 1 | 16 | 16 |
72
  | Gradient Accumulation Steps | 32 | 16 | 1 | 1 |
73
  | Steps | 30 | 24 | 50 | 50 |
74
  | Total Tokens | 62914560 | 100663296 | 419430400 | 838860800 |
@@ -77,9 +86,37 @@ For training data, we generate long contexts by augmenting [SlimPajama](https://
77
  | GPU Type | NVIDIA L40S | NVIDIA L40S | NVIDIA L40S | NVIDIA L40S |
78
  | Minutes to Train (Wall)| 202 | 555 | 61 | 87 |
79
 
80
- **Quants**:
81
- - [GGUF](https://huggingface.co/crusoeai/Llama-3-8B-Instruct-1048k-GGUF)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
82
  - [MLX-4bit](https://huggingface.co/mlx-community/Llama-3-8B-Instruct-1048k-4bit)
 
 
 
83
 
84
  ## The Gradient AI Team
85
 
@@ -99,6 +136,9 @@ Drop an email to [[email protected]](mailto:[email protected])
99
 
100
  [3] https://github.com/jzhang38/EasyContext
101
 
 
 
 
102
 
103
  ----
104
 
 
36
 
37
  # Original Model Card
38
 
39
+ <a href="https://www.gradient.ai" target="_blank"><img src="https://cdn-uploads.huggingface.co/production/uploads/655bb613e8a8971e89944f3e/TSa3V8YpoVagnTYgxiLaO.png" width="200"/></a>
40
+
41
  # Llama-3 8B Gradient Instruct 1048k
42
+
43
+ Join our custom agent and long context (262k-1M+) waitlist: https://forms.gle/L6TDY7dozx8TuoUv7
44
+
45
  Gradient incorporates your data to deploy autonomous assistants that power critical operations across your business. If you're looking to build custom AI models or agents, email us a message [email protected].
46
 
47
  For more info see our [End-to-end development service for custom LLMs and AI systems](https://gradient.ai/development-lab)
48
 
49
+ [Join our Discord](https://discord.com/invite/2QVy2qt2mf)
50
+
51
  This model extends LLama-3 8B's context length from 8k to > 1040K, developed by Gradient, sponsored by compute from [Crusoe Energy](https://huggingface.co/crusoeai). It demonstrates that SOTA LLMs can learn to operate on long context with minimal training by appropriately adjusting RoPE theta. We trained on 830M tokens for this stage, and 1.4B tokens total for all stages, which is < 0.01% of Llama-3's original pre-training data.
52
 
53
+ **Update (5/3): We further fine-tuned our model to strengthen its assistant-like chat ability as well. The NIAH result is updated.**
54
+
55
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585dc9be92bc5f258156bd6/-qaI__83ksClzoJzlqZjq.png)
56
 
57
  **Approach:**
58
 
 
68
 
69
  **Data:**
70
 
71
+ For training data, we generate long contexts by augmenting [SlimPajama](https://huggingface.co/datasets/cerebras/SlimPajama-627B). We also fine-tune on a chat dataset based on UltraChat [4], following a similar recipe for data augmentation to [2].
72
 
73
  **Progressive Training Details:**
74
 
 
77
  | Initialize From | LLaMA-3 8B| 65K | 262K | 524k |
78
  | Sequence Length 2^N | 16 | 18 | 19 | 20 |
79
  | RoPE theta | 15.3 M | 207.1 M | 1.06B | 2.80B |
80
+ | Batch Size | 1 | 1 | 16 | 8 |
81
  | Gradient Accumulation Steps | 32 | 16 | 1 | 1 |
82
  | Steps | 30 | 24 | 50 | 50 |
83
  | Total Tokens | 62914560 | 100663296 | 419430400 | 838860800 |
 
86
  | GPU Type | NVIDIA L40S | NVIDIA L40S | NVIDIA L40S | NVIDIA L40S |
87
  | Minutes to Train (Wall)| 202 | 555 | 61 | 87 |
88
 
89
+
90
+ **Evaluation:**
91
+
92
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6585dc9be92bc5f258156bd6/mWxIGZNi3ejlmeIDWafKu.png)
93
+
94
+ ```
95
+ EVAL_MAX_CONTEXT_LENGTH=1040200
96
+ EVAL_MIN_CONTEXT_LENGTH=100
97
+ EVAL_CONTEXT_INTERVAL=86675
98
+ EVAL_DEPTH_INTERVAL=0.2
99
+ EVAL_RND_NUMBER_DIGITS=8
100
+
101
+ HAYSTACK1:
102
+ EVAL_GENERATOR_TOKENS=25
103
+
104
+ HAYSTACK2:
105
+ EVAL_CONTEXT_INTERVAL=173350
106
+ EVAL_GENERATOR_TOKENS=150000
107
+
108
+ HAYSTACK3:
109
+ EVAL_GENERATOR_TOKENS=925000
110
+ ```
111
+
112
+ All boxes not pictured for Haystack 1 and 3 are 100% accurate. Haystacks 1,2 and 3 are further detailed in this [blog post](https://gradient.ai/blog/the-haystack-matters-for-niah-evals).
113
+
114
+ **Quants:**
115
+ - [GGUF by Crusoe](https://huggingface.co/crusoeai/Llama-3-8B-Instruct-1048k-GGUF). Note that you need to add 128009 as [special token with llama.cpp](https://huggingface.co/gradientai/Llama-3-8B-Instruct-262k/discussions/13).
116
  - [MLX-4bit](https://huggingface.co/mlx-community/Llama-3-8B-Instruct-1048k-4bit)
117
+ - [Ollama](https://ollama.com/library/llama3-gradient)
118
+ - vLLM docker image, recommended to load via `--max-model-len 32768`
119
+ - If you are interested in a hosted version, drop us a mail below.
120
 
121
  ## The Gradient AI Team
122
 
 
136
 
137
  [3] https://github.com/jzhang38/EasyContext
138
 
139
+ [4] Ning Ding, Yulin Chen, Bokai Xu, Yujia Qin, Zhi Zheng, Shengding Hu, Zhiyuan
140
+ Liu, Maosong Sun, and Bowen Zhou. Enhancing chat language models by scaling
141
+ high-quality instructional conversations. arXiv preprint arXiv:2305.14233, 2023.
142
 
143
  ----
144