Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,55 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- dongsheng/DTA-Tool
|
5 |
+
base_model:
|
6 |
+
- meta-llama/Llama-2-13b
|
7 |
+
---
|
8 |
+
|
9 |
+
## Model Description
|
10 |
+
|
11 |
+
<!-- Provide a longer summary of what this model is. -->
|
12 |
+
DTA_llama2_7b is from the paper "[Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation](https://arxiv.org/abs/2501.12432)".
|
13 |
+
It is a large language model capable of invoking tools and can parallel invoke multiple tools within a single round.
|
14 |
+
The tool format it used is similar to OpenAI's Function Call.
|
15 |
+
|
16 |
+
## Uses
|
17 |
+
|
18 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
19 |
+
The related code can be found in our GitHub [repository](https://github.com/Zhudongsheng75/Divide-Then-Aggregate).
|
20 |
+
|
21 |
+
## Training Data
|
22 |
+
|
23 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
24 |
+
|
25 |
+
The training data comes from our specially constructed [DTA-Tool](https://huggingface.co/datasets/dongsheng/DTA-Toolhttps://github.com/OpenBMB/ToolBench), which is derived from [ToolBench](https://github.com/OpenBMB/ToolBench).
|
26 |
+
|
27 |
+
## Evaluation
|
28 |
+
|
29 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
30 |
+
|
31 |
+
### Testing Data
|
32 |
+
|
33 |
+
<!-- This should link to a Dataset Card if possible. -->
|
34 |
+
|
35 |
+
We evaluated the performance of DTA-Llama on [StableToolBench](https://github.com/THUNLP-MT/StableToolBench).
|
36 |
+
|
37 |
+
### Results
|
38 |
+
|
39 |
+

|
40 |
+
|
41 |
+
## Citation
|
42 |
+
|
43 |
+
<!-- If there is a paper or blog post introducing the model, the APA
|
44 |
+
 that should go in this section. -->
|
45 |
+
```bibtex
|
46 |
+
@misc{zhu2025dividethenaggregateefficienttoollearning,
|
47 |
+
title={Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation},
|
48 |
+
author={Dongsheng Zhu and Weixian Shi and Zhengliang Shi and Zhaochun Ren and Shuaiqiang Wang and Lingyong Yan and Dawei Yin},
|
49 |
+
year={2025},
|
50 |
+
eprint={2501.12432},
|
51 |
+
archivePrefix={arXiv},
|
52 |
+
primaryClass={cs.LG},
|
53 |
+
url={https://arxiv.org/abs/2501.12432},
|
54 |
+
}
|
55 |
+
```
|