tlphams commited on
Commit
34aa2e5
1 Parent(s): 76b85d5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +49 -0
README.md CHANGED
@@ -1,3 +1,52 @@
1
  ---
2
  license: cc-by-nc-sa-4.0
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: cc-by-nc-sa-4.0
3
+ pipeline_tag: text-generation
4
+ language:
5
+ - en
6
+ tags:
7
+ - finetuned
8
  ---
9
+
10
+ # Model Card for ZoyLLM-7B-SlimOrca
11
+
12
+ The ZoyLLM-7B-SlimOrca Large Language Model (LLM) is a LoRA-finetuned generative text model with Mistral-7B-v0.1 is the base model.
13
+ Mistral-7B-v0.1 outperforms Llama 2 13B on all benchmarks we tested.
14
+
15
+ ## Model Architecture
16
+
17
+ ZoyLLM-7B-SlimOrca is a transformer model, with the following architecture choices:
18
+ - Grouped-Query Attention
19
+ - Sliding-Window Attention
20
+ - Byte-fallback BPE tokenizer
21
+
22
+ ## Datasets
23
+ - Self-introduction (20 samples)
24
+ - SlimOrca (100k samples random sampled)
25
+ - EverythingLM v3
26
+
27
+ ## Template
28
+ We finetuned the model using a template similar to the dolphin chat template
29
+ ```
30
+ <|im_start|>system
31
+ {system}<|im_end|>
32
+ <|im_start|>user
33
+ {prompt}<|im_end|>
34
+ <|im_start|>assistant
35
+ ```
36
+
37
+ ## Troubleshooting
38
+
39
+ - If you see the following error:
40
+ ```
41
+ KeyError: 'mistral'
42
+ ```
43
+ - Or:
44
+ ```
45
+ NotImplementedError: Cannot copy out of meta tensor; no data!
46
+ ```
47
+
48
+ Ensure you are utilizing a stable version of Transformers, 4.34.0 or newer.
49
+
50
+ ## The Zoy AI Team
51
+
52
+ Pham Tung Lam, Nguyen Duc Nhan.