updated readme
Browse files
README.md
CHANGED
@@ -28,17 +28,25 @@ Evaluating how well the Fine tuned model works as a psychology assistant
|
|
28 |
evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
|
29 |
|
30 |
Models:
|
|
|
31 |
oskaralf/tiny_merged
|
|
|
32 |
TinyLlama (4bit)
|
33 |
- Fine-tuned on FineTome-100k
|
|
|
34 |
Llama3.2_1B_4bit
|
|
|
35 |
Llama 3.2-1B-Instruct (4bit)
|
|
|
36 |
- Fine-tuned on FineTome-100k
|
|
|
37 |
Llama1B_psych_v2_merged
|
|
|
38 |
- Llama 3.2-1B-Instruct (4bit)
|
39 |
- Fine-tuned on samhog/psychology10K
|
40 |
- Lora r=8
|
41 |
Llama1B_psych_v3_merged
|
|
|
42 |
- Llama 3.2-1B-Instruct (4bit)
|
43 |
- Fine-tuned on samhog/psychology10K
|
44 |
- Lora r=16
|
@@ -47,14 +55,26 @@ Llama1B_psych_v3_merged
|
|
47 |
- 4 epochs
|
48 |
|
49 |
To improve:
|
|
|
50 |
Model centric approach
|
|
|
51 |
change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
|
|
|
52 |
Using bigger model
|
|
|
53 |
Training more epochs
|
|
|
54 |
limited due to RAM and time constraint
|
|
|
55 |
Change learning rate
|
56 |
|
57 |
|
|
|
|
|
|
|
|
|
|
|
|
|
58 |
Comparison between different responses on a question:
|
59 |
|
60 |
“I have anxiety, what can I do?”
|
|
|
28 |
evaluating simply on different fine-tuned models how the same phrase performs on different fine-tuned models
|
29 |
|
30 |
Models:
|
31 |
+
|
32 |
oskaralf/tiny_merged
|
33 |
+
|
34 |
TinyLlama (4bit)
|
35 |
- Fine-tuned on FineTome-100k
|
36 |
+
|
37 |
Llama3.2_1B_4bit
|
38 |
+
|
39 |
Llama 3.2-1B-Instruct (4bit)
|
40 |
+
|
41 |
- Fine-tuned on FineTome-100k
|
42 |
+
|
43 |
Llama1B_psych_v2_merged
|
44 |
+
|
45 |
- Llama 3.2-1B-Instruct (4bit)
|
46 |
- Fine-tuned on samhog/psychology10K
|
47 |
- Lora r=8
|
48 |
Llama1B_psych_v3_merged
|
49 |
+
|
50 |
- Llama 3.2-1B-Instruct (4bit)
|
51 |
- Fine-tuned on samhog/psychology10K
|
52 |
- Lora r=16
|
|
|
55 |
- 4 epochs
|
56 |
|
57 |
To improve:
|
58 |
+
|
59 |
Model centric approach
|
60 |
+
|
61 |
change r=16 to higher dimension, for more complex LORA matrices, capturing more complex patterns
|
62 |
+
|
63 |
Using bigger model
|
64 |
+
|
65 |
Training more epochs
|
66 |
+
|
67 |
limited due to RAM and time constraint
|
68 |
+
|
69 |
Change learning rate
|
70 |
|
71 |
|
72 |
+
|
73 |
+
|
74 |
+
|
75 |
+
|
76 |
+
|
77 |
+
|
78 |
Comparison between different responses on a question:
|
79 |
|
80 |
“I have anxiety, what can I do?”
|