kshitizkhanal7's picture
Update README.md
74652c7 verified
---
library_name: transformers
tags:
- climate-change
- flan-t5
- qlora
- instruction-tuning
---
# Model Card for FLAN-T5 Climate Action QLoRA
This is a QLoRA-finetuned version of FLAN-T5 specifically trained for climate action content analysis and generation. The model is optimized for processing and analyzing text related to climate change, sustainability, and environmental policies.
## Model Details
### Model Description
- **Developed by:** Kshitiz Khanal
- **Shared by:** kshitizkhanal7
- **Model type:** Instruction-tuned Language Model with QLoRA fine-tuning
- **Language(s):** English
- **License:** Apache 2.0
- **Finetuned from model:** google/flan-t5-base
### Model Sources
- **Repository:** https://huggingface.co/kshitizkhanal7/flan-t5-climate-qlora
- **Training Data:** FineWeb dataset (climate action filtered)
## Uses
### Direct Use
The model is designed for:
- Analyzing climate policies and initiatives
- Summarizing climate action documents
- Answering questions about climate change and environmental policies
- Evaluating sustainability measures
- Processing climate-related research and reports
### Downstream Use
The model can be integrated into:
- Climate policy analysis tools
- Environmental reporting systems
- Sustainability assessment frameworks
- Climate research applications
- Educational tools about climate change
### Out-of-Scope Use
The model should not be used for:
- Critical policy decisions without human oversight
- Generation of climate misinformation
- Technical climate science research without expert validation
- Commercial deployment without proper testing
- Medical or legal advice
## Bias, Risks, and Limitations
- Limited to climate-related content analysis
- May not perform well on general domain tasks
- Potential biases from web-based training data
- Should not be the sole source for critical decisions
- Performance varies on technical climate science topics
### Recommendations
- Always verify model outputs with authoritative sources
- Use human expert oversight for critical applications
- Consider the model as a supplementary tool, not a replacement for expert knowledge
- Regular evaluation of outputs for potential biases
- Use in conjunction with other data sources for comprehensive analysis
## Training Details
### Training Data
- Source: FineWeb dataset filtered for climate content
- Selection criteria: Climate-related keywords and quality metrics
- Processing: Instruction-style formatting with climate focus
### Training Procedure
#### Preprocessing
- Text cleaning and normalization
- Instruction templates for climate context
- Maximum input length: 512 tokens
- Maximum output length: 128 tokens
#### Training Hyperparameters
- Training regime: QLoRA 4-bit fine-tuning
- Epochs: 3
- Learning rate: 2e-4
- Batch size: 4
- Gradient accumulation steps: 4
- LoRA rank: 16
- LoRA alpha: 32
- Target modules: Query and Value matrices
- LoRA dropout: 0.05
## Environmental Impact
- **Hardware Type:** Single GPU
- **Hours used:** ~4 hours
- **Cloud Provider:** Local
- **Carbon Emitted:** Minimal due to QLoRA efficiency
## Technical Specifications
### Model Architecture and Objective
- Base architecture: FLAN-T5
- Objective: Climate-specific text analysis
- QLoRA adaptation for efficient fine-tuning
- 4-bit quantization for reduced memory usage
### Compute Infrastructure
- Python 3.8+
- PyTorch
- Transformers library
- bitsandbytes for quantization
- PEFT for LoRA implementation
### Hardware
Minimum requirements:
- 16GB GPU memory for inference
- 24GB GPU memory recommended for training
- CPU inference possible but slower
## Citation
If you use this model, please cite:
```bibtex
@misc{khanal2024climate,
title={FLAN-T5 Climate Action QLoRA},
author={Khanal, Kshitiz},
year={2024},
publisher={HuggingFace},
howpublished={\url{https://huggingface.co/kshitizkhanal7/flan-t5-climate-qlora}}
}