Update README.md
Browse files
README.md
CHANGED
@@ -4,10 +4,44 @@ language:
|
|
4 |
- sv
|
5 |
- da
|
6 |
- 'no'
|
7 |
-
- en
|
8 |
pipeline_tag: text-generation
|
9 |
inference:
|
10 |
parameters:
|
11 |
temperature: 0.6
|
|
|
|
|
|
|
|
|
|
|
12 |
---
|
13 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
- sv
|
5 |
- da
|
6 |
- 'no'
|
|
|
7 |
pipeline_tag: text-generation
|
8 |
inference:
|
9 |
parameters:
|
10 |
temperature: 0.6
|
11 |
+
tags:
|
12 |
+
- pytorch
|
13 |
+
- llama
|
14 |
+
- llama-3
|
15 |
+
- ai-sweden
|
16 |
---
|
17 |
+
|
18 |
+
### Use with transformers
|
19 |
+
|
20 |
+
See the snippet below for usage with Transformers:
|
21 |
+
|
22 |
+
```python
|
23 |
+
>>> import transformers
|
24 |
+
>>> import torch
|
25 |
+
|
26 |
+
>>> model_id = "AI-Sweden-Models/Llama-3-8B"
|
27 |
+
|
28 |
+
>>> pipeline = transformers.pipeline(
|
29 |
+
"text-generation",
|
30 |
+
model=model_id,
|
31 |
+
model_kwargs={"torch_dtype": torch.bfloat16},
|
32 |
+
device_map="auto"
|
33 |
+
)
|
34 |
+
>>> pipeline("Hey how are you doing today?")
|
35 |
+
```
|
36 |
+
|
37 |
+
## Training Data
|
38 |
+
|
39 |
+
`AI-Sweden-Models/Llama-3-8B` was trained on a subset from [The nordic pile](https://arxiv.org/abs/2303.17183)
|
40 |
+
|
41 |
+
## Hardware and Software
|
42 |
+
|
43 |
+
**Training Factors** We used custom training libraries, Meta's Research SuperCluster, and production clusters for pretraining. Fine-tuning, annotation, and evaluation were also performed on third-party cloud compute.
|
44 |
+
|
45 |
+
## Benchmarks
|
46 |
+
|
47 |
+
Coming soon.
|