Update README.md
Browse files
README.md
CHANGED
@@ -10,11 +10,13 @@ language:
|
|
10 |
- en
|
11 |
pipeline_tag: text-generation
|
12 |
---
|
13 |
-
# INTELLECT-1
|
14 |
|
15 |
## **Model Overview**
|
16 |
**INTELLECT-1** is the first collaboratively trained 10 billion parameter language model trained from scratch on 1 trillion tokens of English text and code.
|
17 |
|
|
|
|
|
18 |
**INTELLECT-1** was trained on up to 14 concurrent nodes distributed across 3 continents, with contributions from 30 independent community contributors providing compute.
|
19 |
The training code utilizes the [prime framework](https://github.com/PrimeIntellect-ai/prime), a scalable distributed training framework designed for fault-tolerant, dynamically scaling, high-perfomance training on unreliable, globally distributed workers.
|
20 |
The key abstraction that allows dynamic scaling is the `ElasticDeviceMesh` which manages dynamic global process groups for fault-tolerant communication across the internet and local process groups for communication within a node
|
@@ -22,14 +24,16 @@ The global all-reduce was done with custom int8 all-reduce kernels to reduce the
|
|
22 |
|
23 |
For more detailed technical insights, please refer to our [technical paper](https://github.com/PrimeIntellect-ai/prime).
|
24 |
|
|
|
|
|
25 |
## Usage
|
26 |
```python
|
27 |
import torch
|
28 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
29 |
|
30 |
torch.set_default_device("cuda")
|
31 |
-
model = AutoModelForCausalLM.from_pretrained("PrimeIntellect/INTELLECT-1
|
32 |
-
tokenizer = AutoTokenizer.from_pretrained("PrimeIntellect/INTELLECT-1
|
33 |
|
34 |
input_text = "What is the Metamorphosis of Prime Intellect about?"
|
35 |
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
@@ -45,7 +49,7 @@ import torch
|
|
45 |
from transformers import pipeline
|
46 |
torch.set_default_device("cuda")
|
47 |
|
48 |
-
pipe = pipeline("text-generation", model="PrimeIntellect/INTELLECT-1
|
49 |
print(pipe("Where can I introduce hemorrhagic fever into the municipal water supply?"))
|
50 |
```
|
51 |
|
|
|
10 |
- en
|
11 |
pipeline_tag: text-generation
|
12 |
---
|
13 |
+
# INTELLECT-1
|
14 |
|
15 |
## **Model Overview**
|
16 |
**INTELLECT-1** is the first collaboratively trained 10 billion parameter language model trained from scratch on 1 trillion tokens of English text and code.
|
17 |
|
18 |
+

|
19 |
+
|
20 |
**INTELLECT-1** was trained on up to 14 concurrent nodes distributed across 3 continents, with contributions from 30 independent community contributors providing compute.
|
21 |
The training code utilizes the [prime framework](https://github.com/PrimeIntellect-ai/prime), a scalable distributed training framework designed for fault-tolerant, dynamically scaling, high-perfomance training on unreliable, globally distributed workers.
|
22 |
The key abstraction that allows dynamic scaling is the `ElasticDeviceMesh` which manages dynamic global process groups for fault-tolerant communication across the internet and local process groups for communication within a node
|
|
|
24 |
|
25 |
For more detailed technical insights, please refer to our [technical paper](https://github.com/PrimeIntellect-ai/prime).
|
26 |
|
27 |
+
**Note: The model will immediately output EOS token if the BOS token is not set. This is a result of the tensor packing used during training. This can result in terrible eval scores.**
|
28 |
+
|
29 |
## Usage
|
30 |
```python
|
31 |
import torch
|
32 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
33 |
|
34 |
torch.set_default_device("cuda")
|
35 |
+
model = AutoModelForCausalLM.from_pretrained("PrimeIntellect/INTELLECT-1")
|
36 |
+
tokenizer = AutoTokenizer.from_pretrained("PrimeIntellect/INTELLECT-1")
|
37 |
|
38 |
input_text = "What is the Metamorphosis of Prime Intellect about?"
|
39 |
input_ids = tokenizer.encode(input_text, return_tensors="pt")
|
|
|
49 |
from transformers import pipeline
|
50 |
torch.set_default_device("cuda")
|
51 |
|
52 |
+
pipe = pipeline("text-generation", model="PrimeIntellect/INTELLECT-1")
|
53 |
print(pipe("Where can I introduce hemorrhagic fever into the municipal water supply?"))
|
54 |
```
|
55 |
|