oskaralf commited on
Commit
963459b
·
1 Parent(s): 3d30ab4

updated readme

Browse files
Files changed (1) hide show
  1. README.md +20 -0
README.md CHANGED
@@ -28,17 +28,25 @@ Evaluating how well the Fine tuned model works as a psychology assistant
28
  evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
29
 
30
  Models:
 
31
  oskaralf/tiny_merged
 
32
  TinyLlama (4bit)
33
  - Fine-tuned on FineTome-100k
 
34
  Llama3.2_1B_4bit
 
35
  Llama 3.2-1B-Instruct (4bit)
 
36
  - Fine-tuned on FineTome-100k
 
37
  Llama1B_psych_v2_merged
 
38
  - Llama 3.2-1B-Instruct (4bit)
39
  - Fine-tuned on samhog/psychology10K
40
  - Lora r=8
41
  Llama1B_psych_v3_merged
 
42
  - Llama 3.2-1B-Instruct (4bit)
43
  - Fine-tuned on samhog/psychology10K
44
  - Lora r=16
@@ -47,14 +55,26 @@ Llama1B_psych_v3_merged
47
  - 4 epochs
48
 
49
  To improve:
 
50
  Model centric approach
 
51
  change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
 
52
  Using bigger model
 
53
  Training more epochs
 
54
  limited due to RAM and time constraint
 
55
  Change learning rate
56
 
57
 
 
 
 
 
 
 
58
  Comparison between different responses on a question:
59
 
60
  “I have anxiety, what can I do?”
 
28
  evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
29
 
30
  Models:
31
+
32
  oskaralf/tiny_merged
33
+
34
  TinyLlama (4bit)
35
  - Fine-tuned on FineTome-100k
36
+
37
  Llama3.2_1B_4bit
38
+
39
  Llama 3.2-1B-Instruct (4bit)
40
+
41
  - Fine-tuned on FineTome-100k
42
+
43
  Llama1B_psych_v2_merged
44
+
45
  - Llama 3.2-1B-Instruct (4bit)
46
  - Fine-tuned on samhog/psychology10K
47
  - Lora r=8
48
  Llama1B_psych_v3_merged
49
+
50
  - Llama 3.2-1B-Instruct (4bit)
51
  - Fine-tuned on samhog/psychology10K
52
  - Lora r=16
 
55
  - 4 epochs
56
 
57
  To improve:
58
+
59
  Model centric approach
60
+
61
  change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
62
+
63
  Using bigger model
64
+
65
  Training more epochs
66
+
67
  limited due to RAM and time constraint
68
+
69
  Change learning rate
70
 
71
 
72
+
73
+
74
+
75
+
76
+
77
+
78
  Comparison between different responses on a question:
79
 
80
  “I have anxiety, what can I do?”