Spaces:
Sleeping
Sleeping
title: Fine-Tuned BERT Model | |
emoji: 🌍 | |
colorFrom: blue | |
colorTo: purple | |
sdk: docker | |
pinned: true | |
# Fine-Tuned BERT Model for Climate Disinformation Classification | |
## Model Description | |
This is a fine-tuned BERT model trained for the Frugal AI Challenge 2024. The model has been fine-tuned on the climate disinformation dataset to classify text inputs into 8 distinct categories related to climate disinformation. It leverages BERT's pretrained language understanding capabilities and has been optimized for accuracy in this domain. | |
### Intended Use | |
- **Primary intended uses**: Classifying text inputs to detect specific claims of climate disinformation | |
- **Primary intended users**: Researchers, developers, and participants in the Frugal AI Challenge | |
- **Out-of-scope use cases**: Not recommended for tasks outside climate disinformation classification or production-level applications without further evaluation | |
## Training Data | |
The model uses the QuotaClimat/frugalaichallenge-text-train dataset: | |
- Size: ~6000 examples | |
- Split: 80% train, 20% test | |
- 8 categories of climate disinformation claims | |
### Labels | |
0. No relevant claim detected | |
1. Global warming is not happening | |
2. Not caused by humans | |
3. Not bad or beneficial | |
4. Solutions harmful/unnecessary | |
5. Science is unreliable | |
6. Proponents are biased | |
7. Fossil fuels are needed | |
## Performance | |
### Metrics | |
- **Accuracy**: Achieved XX.X% on the test set (replace `XX.X%` with the actual accuracy from your evaluation) | |
- **Environmental Impact**: | |
- Carbon emissions tracked in gCO2eq | |
- Energy consumption tracked in Wh | |
### Model Architecture | |
This model fine-tunes the BERT base architecture (`bert-base-uncased`) for the climate disinformation task. The classifier head includes: | |
- Dense layers | |
- Dropout for regularization | |
- Softmax activation for multi-class classification | |
## Environmental Impact | |
Environmental impact is tracked using CodeCarbon, measuring: | |
- Carbon emissions during inference and training | |
- Energy consumption during inference and training | |
This tracking aligns with the Frugal AI Challenge's commitment to promoting sustainable AI practices. | |
## Limitations | |
- Fine-tuned specifically for climate disinformation; performance on other text classification tasks may degrade | |
- Requires computational resources (e.g., GPU) for efficient inference | |
- Predictions rely on the training dataset's representativeness; may struggle with unseen or out-of-distribution data | |
## Ethical Considerations | |
- Dataset contains sensitive topics related to climate disinformation | |
- Model performance depends on the quality of the dataset and annotation biases | |
- Environmental impact during training and inference is disclosed to encourage awareness of AI's carbon footprint | |
- Users must validate outputs before using in sensitive or high-stakes applications | |
--- | |