timpal0l commited on
Commit
f50b9cf
·
verified ·
1 Parent(s): 38dcdc6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +36 -2
README.md CHANGED
@@ -4,10 +4,44 @@ language:
4
  - sv
5
  - da
6
  - 'no'
7
- - en
8
  pipeline_tag: text-generation
9
  inference:
10
  parameters:
11
  temperature: 0.6
 
 
 
 
 
12
  ---
13
- # Model Card
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  - sv
5
  - da
6
  - 'no'
 
7
  pipeline_tag: text-generation
8
  inference:
9
  parameters:
10
  temperature: 0.6
11
+ tags:
12
+ - pytorch
13
+ - llama
14
+ - llama-3
15
+ - ai-sweden
16
  ---
17
+
18
+ ### Use with transformers
19
+
20
+ See the snippet below for usage with Transformers:
21
+
22
+ ```python
23
+ >>> import transformers
24
+ >>> import torch
25
+
26
+ >>> model_id = "AI-Sweden-Models/Llama-3-8B"
27
+
28
+ >>> pipeline = transformers.pipeline(
29
+ "text-generation",
30
+ model=model_id,
31
+ model_kwargs={"torch_dtype": torch.bfloat16},
32
+ device_map="auto"
33
+ )
34
+ >>> pipeline("Hey how are you doing today?")
35
+ ```
36
+
37
+ ## Training Data
38
+
39
+ `AI-Sweden-Models/Llama-3-8B` was trained on a subset from [The nordic pile](https://arxiv.org/abs/2303.17183)
40
+
41
+ ## Hardware and Software
42
+
43
+ **Training Factors** We used custom training libraries, Meta's Research SuperCluster, and production clusters for pretraining. Fine-tuning, annotation, and evaluation were also performed on third-party cloud compute.
44
+
45
+ ## Benchmarks
46
+
47
+ Coming soon.