PyTorch
llama
dongsheng commited on
Commit
e8d7e9e
·
verified ·
1 Parent(s): a62c207

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -0
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - dongsheng/DTA-Tool
5
+ base_model:
6
+ - meta-llama/Llama-2-13b
7
+ ---
8
+
9
+ ## Model Description
10
+
11
+ <!-- Provide a longer summary of what this model is. -->
12
+ DTA_llama2_7b is from the paper "[Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation](https://arxiv.org/abs/2501.12432)".
13
+ It is a large language model capable of invoking tools and can parallel invoke multiple tools within a single round.
14
+ The tool format it used is similar to OpenAI's Function Call.
15
+
16
+ ## Uses
17
+
18
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
19
+ The related code can be found in our GitHub [repository](https://github.com/Zhudongsheng75/Divide-Then-Aggregate).
20
+
21
+ ## Training Data
22
+
23
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
24
+
25
+ The training data comes from our specially constructed [DTA-Tool](https://huggingface.co/datasets/dongsheng/DTA-Toolhttps://github.com/OpenBMB/ToolBench), which is derived from [ToolBench](https://github.com/OpenBMB/ToolBench).
26
+
27
+ ## Evaluation
28
+
29
+ <!-- This section describes the evaluation protocols and provides the results. -->
30
+
31
+ ### Testing Data
32
+
33
+ <!-- This should link to a Dataset Card if possible. -->
34
+
35
+ We evaluated the performance of DTA-Llama on [StableToolBench](https://github.com/THUNLP-MT/StableToolBench).
36
+
37
+ ### Results
38
+
39
+ ![result](result.png)
40
+
41
+ ## Citation
42
+
43
+ <!-- If there is a paper or blog post introducing the model, the APA
44
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/630da0fae57da204209411d3/ViBSn34pV-4LWJkIpUvSr.png) that should go in this section. -->
45
+ ```bibtex
46
+ @misc{zhu2025dividethenaggregateefficienttoollearning,
47
+ title={Divide-Then-Aggregate: An Efficient Tool Learning Method via Parallel Tool Invocation},
48
+ author={Dongsheng Zhu and Weixian Shi and Zhengliang Shi and Zhaochun Ren and Shuaiqiang Wang and Lingyong Yan and Dawei Yin},
49
+ year={2025},
50
+ eprint={2501.12432},
51
+ archivePrefix={arXiv},
52
+ primaryClass={cs.LG},
53
+ url={https://arxiv.org/abs/2501.12432},
54
+ }
55
+ ```