Muning9999
commited on
Commit
•
7ea28be
1
Parent(s):
b7f13df
Update README.md
Browse files
README.md
CHANGED
@@ -8,7 +8,7 @@ base_model: Qwen/Qwen1.5-4B-Chat
|
|
8 |
# Hammer-4b Function Calling Model
|
9 |
|
10 |
## Introduction
|
11 |
-
Hammer
|
12 |
|
13 |
## Model Details
|
14 |
Hammer-4b is a finetuned model built upon [Qwen1.5-4B-Chat](https://huggingface.co/Qwen/Qwen1.5-4B-Chat). It's trained using the [APIGen Function Calling Datasets](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k) containing 60,000 samples, supplemented by [7,500 irrelevance detection data](https://huggingface.co/datasets/MadeAgents/XLAM-7.5k-Irrelevance) we generated. Employing innovative training techniques like function masking, function shuffling, and prompt optimization, Hammer-4b has achieved exceptional performances across numerous benchmarks including [Berkley Function Calling Leaderboard](https://gorilla.cs.berkeley.edu/leaderboard.html), [API-Bank](https://arxiv.org/abs/2304.08244), [Tool-Alpaca](https://arxiv.org/abs/2306.05301), [Nexus Raven](https://github.com/nexusflowai/NexusRaven-V2) and [Seal-Tools](https://arxiv.org/abs/2405.08355).
|
@@ -17,7 +17,7 @@ Hammer-4b is a finetuned model built upon [Qwen1.5-4B-Chat](https://huggingface.
|
|
17 |
Thanks so much for your attention, a report with all the technical details leading to our models will be published soon.
|
18 |
|
19 |
## Evaluation
|
20 |
-
First, we evaluate Hammer-
|
21 |
|
22 |
<div style="text-align: center;">
|
23 |
<img src="figures_bfcl.PNG" alt="overview" width="1480" style="margin: auto;">
|
|
|
8 |
# Hammer-4b Function Calling Model
|
9 |
|
10 |
## Introduction
|
11 |
+
**Hammer** is a series of cutting-edge Large Language Models (LLMs) crafted to boost the critical capability of AI agents: function calling. Differing from existing models focusing on training data refinement, Hammer optimizes performance primarily through advanced training techniques. Focusing on on-device applications, we release a number of models from 1.5B, 4B to 7B parameters.
|
12 |
|
13 |
## Model Details
|
14 |
Hammer-4b is a finetuned model built upon [Qwen1.5-4B-Chat](https://huggingface.co/Qwen/Qwen1.5-4B-Chat). It's trained using the [APIGen Function Calling Datasets](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k) containing 60,000 samples, supplemented by [7,500 irrelevance detection data](https://huggingface.co/datasets/MadeAgents/XLAM-7.5k-Irrelevance) we generated. Employing innovative training techniques like function masking, function shuffling, and prompt optimization, Hammer-4b has achieved exceptional performances across numerous benchmarks including [Berkley Function Calling Leaderboard](https://gorilla.cs.berkeley.edu/leaderboard.html), [API-Bank](https://arxiv.org/abs/2304.08244), [Tool-Alpaca](https://arxiv.org/abs/2306.05301), [Nexus Raven](https://github.com/nexusflowai/NexusRaven-V2) and [Seal-Tools](https://arxiv.org/abs/2405.08355).
|
|
|
17 |
Thanks so much for your attention, a report with all the technical details leading to our models will be published soon.
|
18 |
|
19 |
## Evaluation
|
20 |
+
First, we evaluate Hammer-7b on the Berkeley Function-Calling Leaderboard (BFCL):
|
21 |
|
22 |
<div style="text-align: center;">
|
23 |
<img src="figures_bfcl.PNG" alt="overview" width="1480" style="margin: auto;">
|