File size: 2,459 Bytes
e5dd0fa |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 |
# Model Card for climategpt/climategpt-7b-fsg
- This model is the 7B parameter from-scratch general ("fsg") variant of the ClimateGPT model release.
## Overview
- **Developed by:** AppTek, Eqtylab, Erasmus AI
- **Model type:** decoder-only Transformer
- **Language(s) (NLP):** natively supported: English; supported via cascaded MT on web interface: Arabic, Bangla, Chinese (simplified), Dutch, Finnougoric, French, Germanic, Greek, Hebrew, Indonesian, Japenese, Korean, Lithuanian, Pashto, Persian, Portuguese, Russian, Spanish, Thai, Turkish, Vietnamese,
- **License:** TO BE ADDED
- **Repository:** https://huggingface.co/climategpt/climategpt-7b-fsg
- **Paper:** TO BE ADDED
- **Demo:** TO BE ADDED
## Uses
- This model is intended to be directly used as a question answering model that is specialized in the climate domain.
- The model is aimed at providing useful feedback for decision makers, scientists and jounalists involved in climate discussions.
- The model can also be used as a starting point for interested developers for further finetuning.
- The model is NOT intended to be a general-purpose chatbot (although it has chat capabilities).
- For the full system including cascaded MT, RAG, etc., we recommend the user to go to our demo website: TO BE ADDED.
- For hands-on finetuning deployment and inference, we recommend the user to directly use the Huggingface helpers.
- For in-depth model conversion and finetuning, we recommend the user to use https://github.com/epfLLM/Megatron-LLM/.
- **Despite the efforts from the development team to elimite them, as every other chat-capable LLMs, this model may generate biased, offensive, inaccurate responses.**
## How to Get Started with the Model
After downloading the HF formatted model, the HF helpers should work out-of-the-box.
It is also possible to evaluate the model with https://github.com/EleutherAI/lm-evaluation-harness by plugging in the model identifier ```--model_args pretrained=climategpt/climategpt-7b-fsg```.
## Training
- For pretraining, a 300B-token dataset with an emphasis on the climate domain is prepared and used.
- For instruction finetuning, about 1.1B instruction-finetuning tokens (both in the climate domain but also general domain) are used.
## Environmental Impact
- **Hardware Type:** H100
- **Hours used:** 30720 hrs
- **Cloud Provider:** TO BE ADDED
- **Compute Region:** TO BE ADDED
- **Carbon Emitted:** TO BE ADDED
## Citation
**BibTeX:** TO BE ADDED |