Update README.md
Browse files
README.md
CHANGED
@@ -20,4 +20,17 @@ The only difference from Llama-3.2-1B-chatml-tool-v1 is that it uses AlternateTo
|
|
20 |
|
21 |
In the case of the existing tool-AlternateTokenizer, the <tool_call> tag was not properly generated before the function call, but in v2, it was observed that it performed well when trained with the general AlternateTokenizer.
|
22 |
|
23 |
-
need to check whether this phenomenon is repeated in larger models (3B, 8B).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
20 |
|
21 |
In the case of the existing tool-AlternateTokenizer, the <tool_call> tag was not properly generated before the function call, but in v2, it was observed that it performed well when trained with the general AlternateTokenizer.
|
22 |
|
23 |
+
need to check whether this phenomenon is repeated in larger models (3B, 8B).
|
24 |
+
|
25 |
+
|
26 |
+
|
27 |
+
## Model Performance Comparison (BFCL)
|
28 |
+
|
29 |
+
| task name | minpeter/Llama-3.2-1B-chatml-tool-v2 | meta-llama/Llama-3.2-1B-Instruct |
|
30 |
+
|-----------------|-----------------------------------|-----------------------------------|
|
31 |
+
| parallel_multiple | 0.000 | 0.025 |
|
32 |
+
| parallel | 0.000 | 0.035 |
|
33 |
+
| simple | 0.72 | 0.215 |
|
34 |
+
| multiple | 0.695 | 0.17 |
|
35 |
+
|
36 |
+
*Parallel calls are not taken into account. 0 points are expected. We plan to fix this in v3.
|