passing2961
commited on
Commit
β’
c153553
1
Parent(s):
74743db
Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,56 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-nc-3.0
|
3 |
+
datasets:
|
4 |
+
- passing2961/stark-summary
|
5 |
+
language:
|
6 |
+
- en
|
7 |
+
base_model:
|
8 |
+
- meta-llama/Llama-3.2-1B-Instruct
|
9 |
+
tags:
|
10 |
+
- conversational ai
|
11 |
+
- conversation summarization
|
12 |
+
---
|
13 |
+
|
14 |
+
# Ultron-Summarizer-1B Model Card
|
15 |
+
|
16 |
+
[π Homepage](https://stark-dataset.github.io/) | [π» Github](https://github.com/passing2961/Stark) | [π Arxiv](https://arxiv.org/abs/2407.03958) | [π PDF](https://arxiv.org/pdf/2407.03958)
|
17 |
+
|
18 |
+
## List of Provided Model Series
|
19 |
+
- **Ultron-Summarizer-Series:** [π€ Ultron-Summarizer-1B](https://huggingface.co/passing2961/Ultron-Summarizer-1B) | [π€ Ultron-Summarizer-3B](https://huggingface.co/passing2961/Ultron-Summarizer-3B) | [π€ Ultron-Summarizer-8B](https://huggingface.co/passing2961/Ultron-Summarizer-8B)
|
20 |
+
- **Ultron 7B**: [π€ Ultron-7B](https://huggingface.co/passing2961/Ultron-7B)
|
21 |
+
|
22 |
+
> π¨ Disclaimer: All models and datasets are intended for research purposes only.
|
23 |
+
|
24 |
+
## Model Description
|
25 |
+
- **Repository:** [Code](https://github.com/passing2961/Thanos)
|
26 |
+
- **Paper:** Thanos: Enhancing Conversational Agents with Skill-of-Mind-Infused Large Language Model
|
27 |
+
- **Point of Contact:** [Young-Jun Lee](mailto:[email protected])
|
28 |
+
|
29 |
+
## Model Details
|
30 |
+
- **Model**: Ultron-Summarizer-1B is a fully open-source, conversation summarize that generate the summarization given the long-term conversation.
|
31 |
+
- **Date**: Ultron-Summarizer-1B was trained in 2024.
|
32 |
+
- **Training Dataset**: [Stark-Summary](https://huggingface.co/datasets/passing2961/stark-summary)
|
33 |
+
- **Architecture**: Ultron-Summarizer-1B was trained on top of [LLaMA-3.2-1B](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct).
|
34 |
+
|
35 |
+
## How to Use
|
36 |
+
|
37 |
+
## License and Recommendations
|
38 |
+
|
39 |
+
π¨ Ultron-Summarizer-1B is intended to be used for research purposes only.
|
40 |
+
|
41 |
+
## Acknowledgement
|
42 |
+
|
43 |
+
This work was supported by a grant of the KAIST-KT joint research project through AI Tech Lab, Institute of convergence Technology, funded by KT [Project No. G01230605, Development of Task-oriented Persona-based Dialogue Generation Combining Multi-modal Interaction and Knowledge Modeling].
|
44 |
+
|
45 |
+
## Citation
|
46 |
+
|
47 |
+
If you find the resources in this repository useful, please cite our work:
|
48 |
+
|
49 |
+
```
|
50 |
+
@article{lee2024stark,
|
51 |
+
title={Stark: Social Long-Term Multi-Modal Conversation with Persona Commonsense Knowledge},
|
52 |
+
author={Lee, Young-Jun and Lee, Dokyong and Youn, Junyoung and Oh, Kyeongjin and Ko, Byungsoo and Hyeon, Jonghwan and Choi, Ho-Jin},
|
53 |
+
journal={arXiv preprint arXiv:2407.03958},
|
54 |
+
year={2024}
|
55 |
+
}
|
56 |
+
```
|