metadata
library_name: transformers
tags:
- climate-change
- flan-t5
- qlora
- instruction-tuning
Model Card for FLAN-T5 Climate Action QLoRA
This is a QLoRA-finetuned version of FLAN-T5 specifically trained for climate action content analysis and generation. The model is optimized for processing and analyzing text related to climate change, sustainability, and environmental policies.
Model Details
Model Description
- Developed by: Kshitiz Khanal
- Shared by: kshitizkhanal7
- Model type: Instruction-tuned Language Model with QLoRA fine-tuning
- Language(s): English
- License: Apache 2.0
- Finetuned from model: google/flan-t5-base
Model Sources
- Repository: https://huggingface.co/kshitizkhanal7/flan-t5-climate-qlora
- Training Data: FineWeb dataset (climate action filtered)
Uses
Direct Use
The model is designed for:
- Analyzing climate policies and initiatives
- Summarizing climate action documents
- Answering questions about climate change and environmental policies
- Evaluating sustainability measures
- Processing climate-related research and reports
Downstream Use
The model can be integrated into:
- Climate policy analysis tools
- Environmental reporting systems
- Sustainability assessment frameworks
- Climate research applications
- Educational tools about climate change
Out-of-Scope Use
The model should not be used for:
- Critical policy decisions without human oversight
- Generation of climate misinformation
- Technical climate science research without expert validation
- Commercial deployment without proper testing
- Medical or legal advice
Bias, Risks, and Limitations
- Limited to climate-related content analysis
- May not perform well on general domain tasks
- Potential biases from web-based training data
- Should not be the sole source for critical decisions
- Performance varies on technical climate science topics
Recommendations
- Always verify model outputs with authoritative sources
- Use human expert oversight for critical applications
- Consider the model as a supplementary tool, not a replacement for expert knowledge
- Regular evaluation of outputs for potential biases
- Use in conjunction with other data sources for comprehensive analysis
Training Details
Training Data
- Source: FineWeb dataset filtered for climate content
- Selection criteria: Climate-related keywords and quality metrics
- Processing: Instruction-style formatting with climate focus
Training Procedure
Preprocessing
- Text cleaning and normalization
- Instruction templates for climate context
- Maximum input length: 512 tokens
- Maximum output length: 128 tokens
Training Hyperparameters
- Training regime: QLoRA 4-bit fine-tuning
- Epochs: 3
- Learning rate: 2e-4
- Batch size: 4
- Gradient accumulation steps: 4
- LoRA rank: 16
- LoRA alpha: 32
- Target modules: Query and Value matrices
- LoRA dropout: 0.05
Environmental Impact
- Hardware Type: Single GPU
- Hours used: ~4 hours
- Cloud Provider: Local
- Carbon Emitted: Minimal due to QLoRA efficiency
Technical Specifications
Model Architecture and Objective
- Base architecture: FLAN-T5
- Objective: Climate-specific text analysis
- QLoRA adaptation for efficient fine-tuning
- 4-bit quantization for reduced memory usage
Compute Infrastructure
- Python 3.8+
- PyTorch
- Transformers library
- bitsandbytes for quantization
- PEFT for LoRA implementation
Hardware
Minimum requirements:
- 16GB GPU memory for inference
- 24GB GPU memory recommended for training
- CPU inference possible but slower
Citation
If you use this model, please cite:
@misc{khanal2024climate,
title={FLAN-T5 Climate Action QLoRA},
author={Khanal, Kshitiz},
year={2024},
publisher={HuggingFace},
howpublished={\url{https://huggingface.co/kshitizkhanal7/flan-t5-climate-qlora}}
}