Safetensors
qwen2
Muning9999 commited on
Commit
7ea28be
1 Parent(s): b7f13df

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -8,7 +8,7 @@ base_model: Qwen/Qwen1.5-4B-Chat
8
  # Hammer-4b Function Calling Model
9
 
10
  ## Introduction
11
- Hammer-4b is a cutting-edge Large Language Model (LLM) crafted to boost the critical capability of AI agents: function calling. Differing from existing models focusing on traning data refinement, Hammer-4b optimizes performance primarily through advanced training techniques.
12
 
13
  ## Model Details
14
  Hammer-4b is a finetuned model built upon [Qwen1.5-4B-Chat](https://huggingface.co/Qwen/Qwen1.5-4B-Chat). It's trained using the [APIGen Function Calling Datasets](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k) containing 60,000 samples, supplemented by [7,500 irrelevance detection data](https://huggingface.co/datasets/MadeAgents/XLAM-7.5k-Irrelevance) we generated. Employing innovative training techniques like function masking, function shuffling, and prompt optimization, Hammer-4b has achieved exceptional performances across numerous benchmarks including [Berkley Function Calling Leaderboard](https://gorilla.cs.berkeley.edu/leaderboard.html), [API-Bank](https://arxiv.org/abs/2304.08244), [Tool-Alpaca](https://arxiv.org/abs/2306.05301), [Nexus Raven](https://github.com/nexusflowai/NexusRaven-V2) and [Seal-Tools](https://arxiv.org/abs/2405.08355).
@@ -17,7 +17,7 @@ Hammer-4b is a finetuned model built upon [Qwen1.5-4B-Chat](https://huggingface.
17
  Thanks so much for your attention, a report with all the technical details leading to our models will be published soon.
18
 
19
  ## Evaluation
20
- First, we evaluate Hammer-4b on the Berkeley Function-Calling Leaderboard (BFCL):
21
 
22
  <div style="text-align: center;">
23
  <img src="figures_bfcl.PNG" alt="overview" width="1480" style="margin: auto;">
 
8
  # Hammer-4b Function Calling Model
9
 
10
  ## Introduction
11
+ **Hammer** is a series of cutting-edge Large Language Models (LLMs) crafted to boost the critical capability of AI agents: function calling. Differing from existing models focusing on training data refinement, Hammer optimizes performance primarily through advanced training techniques. Focusing on on-device applications, we release a number of models from 1.5B, 4B to 7B parameters.
12
 
13
  ## Model Details
14
  Hammer-4b is a finetuned model built upon [Qwen1.5-4B-Chat](https://huggingface.co/Qwen/Qwen1.5-4B-Chat). It's trained using the [APIGen Function Calling Datasets](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k) containing 60,000 samples, supplemented by [7,500 irrelevance detection data](https://huggingface.co/datasets/MadeAgents/XLAM-7.5k-Irrelevance) we generated. Employing innovative training techniques like function masking, function shuffling, and prompt optimization, Hammer-4b has achieved exceptional performances across numerous benchmarks including [Berkley Function Calling Leaderboard](https://gorilla.cs.berkeley.edu/leaderboard.html), [API-Bank](https://arxiv.org/abs/2304.08244), [Tool-Alpaca](https://arxiv.org/abs/2306.05301), [Nexus Raven](https://github.com/nexusflowai/NexusRaven-V2) and [Seal-Tools](https://arxiv.org/abs/2405.08355).
 
17
  Thanks so much for your attention, a report with all the technical details leading to our models will be published soon.
18
 
19
  ## Evaluation
20
+ First, we evaluate Hammer-7b on the Berkeley Function-Calling Leaderboard (BFCL):
21
 
22
  <div style="text-align: center;">
23
  <img src="figures_bfcl.PNG" alt="overview" width="1480" style="margin: auto;">