oskaralf commited on
Commit
5b0d177
·
1 Parent(s): 963459b

updated readme

Browse files
Files changed (1) hide show
  1. README.md +20 -14
README.md CHANGED
@@ -12,18 +12,28 @@ pinned: false
12
  In this lab, different LLM's were trained through Google Colab. We mainly explored Llama-1B-Instruct, through different datasets, aiming to finetune the model into acting as a psychologist.
13
 
14
  Ground models evaluated:
 
15
  TinyLlama
16
- smaller, faster, with around 1B parameters
17
- not so good for sophisticated answers
 
 
18
  Llama3.2 _1B_Instruct
 
19
  Llama3.2 _3B_Instruct
20
 
 
 
21
  Data sets used (from Huggingface)
22
- mlabonne/FineTome-100k
23
- wassimm/PsycologyDataset
24
- samhog/psychology-10k
 
 
 
25
 
26
  Evaluation method
 
27
  Evaluating how well the Fine tuned model works as a psychology assistant
28
  evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
29
 
@@ -58,15 +68,11 @@ To improve:
58
 
59
  Model centric approach
60
 
61
- change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
62
-
63
- Using bigger model
64
-
65
- Training more epochs
66
-
67
- limited due to RAM and time constraint
68
-
69
- Change learning rate
70
 
71
 
72
 
 
12
  In this lab, different LLM's were trained through Google Colab. We mainly explored Llama-1B-Instruct, through different datasets, aiming to finetune the model into acting as a psychologist.
13
 
14
  Ground models evaluated:
15
+
16
  TinyLlama
17
+
18
+ - smaller, faster, with around 1B parameters
19
+ - not so good for sophisticated answers
20
+
21
  Llama3.2 _1B_Instruct
22
+
23
  Llama3.2 _3B_Instruct
24
 
25
+
26
+
27
  Data sets used (from Huggingface)
28
+
29
+ - mlabonne/FineTome-100k
30
+ - wassimm/PsycologyDataset
31
+ - samhog/psychology-10k
32
+
33
+
34
 
35
  Evaluation method
36
+
37
  Evaluating how well the Fine tuned model works as a psychology assistant
38
  evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
39
 
 
68
 
69
  Model centric approach
70
 
71
+ - change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
72
+ - Using bigger model
73
+ - Training more epochs
74
+ - limited due to RAM and time constraint
75
+ - Change learning rate
 
 
 
 
76
 
77
 
78