PEFT
Safetensors
Transformers
text-generation-inference
unsloth
llama
trl
mpasila commited on
Commit
74182c7
1 Parent(s): 7c1f721

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -1
README.md CHANGED
@@ -2,6 +2,12 @@
2
  base_model: LumiOpen/Viking-7B
3
  language:
4
  - en
 
 
 
 
 
 
5
  license: apache-2.0
6
  tags:
7
  - text-generation-inference
@@ -9,7 +15,17 @@ tags:
9
  - unsloth
10
  - llama
11
  - trl
 
 
 
12
  ---
 
 
 
 
 
 
 
13
 
14
  # Uploaded model
15
 
@@ -19,4 +35,4 @@ tags:
19
 
20
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
2
  base_model: LumiOpen/Viking-7B
3
  language:
4
  - en
5
+ - fi
6
+ - sv
7
+ - 'no'
8
+ - da
9
+ - is
10
+ - nn
11
  license: apache-2.0
12
  tags:
13
  - text-generation-inference
 
15
  - unsloth
16
  - llama
17
  - trl
18
+ datasets:
19
+ - Gryphe/Sonnet3.5-SlimOrcaDedupCleaned
20
+ - mpasila/Sonnet3.5-SlimOrcaDedupCleaned-4k-context
21
  ---
22
+ This is a test model because the previous attempt failed.
23
+
24
+ Prompt format is: ChatML
25
+
26
+ Merged model: [mpasila/Viking-SlimSonnet-v0.2-7B](https://huggingface.co/mpasila/Viking-SlimSonnet-v0.2-7B)
27
+
28
+ Trained with regular LoRA (not quantized/QLoRA) and LoRA rank was 128 and Alpha set to 32. Trained for 5000 steps.
29
 
30
  # Uploaded model
31
 
 
35
 
36
  This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
37
 
38
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)