upload model
Browse files- README.md +51 -0
- adapter_config.json +31 -0
- adapter_model.bin +3 -0
- llasmol.png +0 -0
README.md
CHANGED
@@ -1,3 +1,54 @@
|
|
1 |
---
|
|
|
|
|
2 |
license: cc-by-4.0
|
|
|
|
|
|
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
language:
|
3 |
+
- en
|
4 |
license: cc-by-4.0
|
5 |
+
tags:
|
6 |
+
- instruction tuning
|
7 |
+
- chemistry
|
8 |
+
- molecule
|
9 |
+
- small molecule
|
10 |
---
|
11 |
+
|
12 |
+
<p align="center">
|
13 |
+
<img style="width: 20%;" src="llasmol.png">
|
14 |
+
</p>
|
15 |
+
|
16 |
+
<h1 align="center"> LlaSMol </h1>
|
17 |
+
<h3 align="center"> LlaSMol-CodeLlama-7B </h3>
|
18 |
+
|
19 |
+
**Paper**: [LlaSMol: Advancing Large Language Models for Chemistry with a Large-Scale, Comprehensive, High-Quality Instruction Tuning Dataset](https://arxiv.org/abs/2402.09391)
|
20 |
+
|
21 |
+
**Page**: [https://osu-nlp-group.github.io/LlaSMol](https://osu-nlp-group.github.io/LlaSMol)
|
22 |
+
|
23 |
+
**Code**: [https://github.com/OSU-NLP-Group/LlaSMol](https://github.com/OSU-NLP-Group/LlaSMol)
|
24 |
+
|
25 |
+
**Models**:
|
26 |
+
- LlaSMol-Galactica-6.7B: [https://huggingface.co/osunlp/LlaSMol-Galactica-6.7B](https://huggingface.co/osunlp/LlaSMol-Galactica-6.7B)
|
27 |
+
- LlaSMol-Llama2-7B: [https://huggingface.co/osunlp/LlaSMol-Llama2-7B](https://huggingface.co/osunlp/LlaSMol-Llama2-7B)
|
28 |
+
- LlaSMol-CodeLlama-7B: [https://huggingface.co/osunlp/LlaSMol-CodeLlama-7B](https://huggingface.co/osunlp/LlaSMol-CodeLlama-7B)
|
29 |
+
- LlaSMol-Mistral-7B: [https://huggingface.co/osunlp/LlaSMol-Mistral-7B](https://huggingface.co/osunlp/LlaSMol-Mistral-7B)
|
30 |
+
|
31 |
+
LlaSMol-CodeLlama-7B is an LLM for chemistry. It is based on [codellama/CodeLlama-7b-hf](https://huggingface.co/codellama/CodeLlama-7b-hf) and tuned on our [SMolInstruct](https://huggingface.co/datasets/osunlp/SMolInstruct) dataset with LoRA. This repo contains the weight of the low-rank adapter.
|
32 |
+
|
33 |
+
## ⚔️ Usage
|
34 |
+
|
35 |
+
For instructions to run the model, please refer to our [repository](https://github.com/OSU-NLP-Group/LlaSMol).
|
36 |
+
|
37 |
+
|
38 |
+
## 🚨 Limitations
|
39 |
+
|
40 |
+
While the model is carefully trained, we do not guarantee its effectiveness. The model may output incorrect or inaccurate information. Please use it at your own risk.
|
41 |
+
|
42 |
+
Additionally, the model is built as a mature product but solely for research purpose. It may generate harmful or biased information. We emphatically urge all users to adhere to the highest ethical standards when using the model, including maintaining fairness, transparency, and responsibility in their research. Any usage of the dataset that may lead to harm or pose a detriment to society is strictly **forbidden**.
|
43 |
+
|
44 |
+
## 📚 Citation
|
45 |
+
If our paper or related resources prove valuable to your research, we kindly ask for citation. Please feel free to contact us with any inquiries.
|
46 |
+
|
47 |
+
```
|
48 |
+
@article{yu2024llasmol,
|
49 |
+
title={LlaSMol: Advancing Large Language Models for Chemistry with a Large-Scale, Comprehensive, High-Quality Instruction Tuning Dataset},
|
50 |
+
author={Botao Yu and Frazier N. Baker and Ziqi Chen and Xia Ning and Huan Sun},
|
51 |
+
journal={arXiv preprint arXiv:2402.09391},
|
52 |
+
year={2024}
|
53 |
+
}
|
54 |
+
```
|
adapter_config.json
ADDED
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"alpha_pattern": {},
|
3 |
+
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "codellama/CodeLlama-7b-hf",
|
5 |
+
"bias": "none",
|
6 |
+
"fan_in_fan_out": false,
|
7 |
+
"inference_mode": true,
|
8 |
+
"init_lora_weights": true,
|
9 |
+
"layers_pattern": null,
|
10 |
+
"layers_to_transform": null,
|
11 |
+
"loftq_config": {},
|
12 |
+
"lora_alpha": 16,
|
13 |
+
"lora_dropout": 0.05,
|
14 |
+
"megatron_config": null,
|
15 |
+
"megatron_core": "megatron.core",
|
16 |
+
"modules_to_save": [],
|
17 |
+
"peft_type": "LORA",
|
18 |
+
"r": 16,
|
19 |
+
"rank_pattern": {},
|
20 |
+
"revision": null,
|
21 |
+
"target_modules": [
|
22 |
+
"gate_proj",
|
23 |
+
"o_proj",
|
24 |
+
"k_proj",
|
25 |
+
"down_proj",
|
26 |
+
"up_proj",
|
27 |
+
"v_proj",
|
28 |
+
"q_proj"
|
29 |
+
],
|
30 |
+
"task_type": "CAUSAL_LM"
|
31 |
+
}
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a508acfa18f721ed32271864620ed7e7b81bcccb0113414678f7923ec6a5d900
|
3 |
+
size 160069389
|
llasmol.png
ADDED