Spaces:
Sleeping
Sleeping
File size: 2,897 Bytes
6fb5d57 a02d840 6fb5d57 a02d840 6fb5d57 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 70f5f26 a02d840 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 |
---
title: Fine-Tuned BERT Model
emoji: 🌍
colorFrom: blue
colorTo: purple
sdk: docker
pinned: true
---
# Fine-Tuned BERT Model for Climate Disinformation Classification
## Model Description
This is a fine-tuned BERT model trained for the Frugal AI Challenge 2024. The model has been fine-tuned on the climate disinformation dataset to classify text inputs into 8 distinct categories related to climate disinformation. It leverages BERT's pretrained language understanding capabilities and has been optimized for accuracy in this domain.
### Intended Use
- **Primary intended uses**: Classifying text inputs to detect specific claims of climate disinformation
- **Primary intended users**: Researchers, developers, and participants in the Frugal AI Challenge
- **Out-of-scope use cases**: Not recommended for tasks outside climate disinformation classification or production-level applications without further evaluation
## Training Data
The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
- Size: ~6000 examples
- Split: 80% train, 20% test
- 8 categories of climate disinformation claims
### Labels
0. No relevant claim detected
1. Global warming is not happening
2. Not caused by humans
3. Not bad or beneficial
4. Solutions harmful/unnecessary
5. Science is unreliable
6. Proponents are biased
7. Fossil fuels are needed
## Performance
### Metrics
- **Accuracy**: Achieved XX.X% on the test set (replace `XX.X%` with the actual accuracy from your evaluation)
- **Environmental Impact**:
- Carbon emissions tracked in gCO2eq
- Energy consumption tracked in Wh
### Model Architecture
This model fine-tunes the BERT base architecture (`bert-base-uncased`) for the climate disinformation task. The classifier head includes:
- Dense layers
- Dropout for regularization
- Softmax activation for multi-class classification
## Environmental Impact
Environmental impact is tracked using CodeCarbon, measuring:
- Carbon emissions during inference and training
- Energy consumption during inference and training
This tracking aligns with the Frugal AI Challenge's commitment to promoting sustainable AI practices.
## Limitations
- Fine-tuned specifically for climate disinformation; performance on other text classification tasks may degrade
- Requires computational resources (e.g., GPU) for efficient inference
- Predictions rely on the training dataset's representativeness; may struggle with unseen or out-of-distribution data
## Ethical Considerations
- Dataset contains sensitive topics related to climate disinformation
- Model performance depends on the quality of the dataset and annotation biases
- Environmental impact during training and inference is disclosed to encourage awareness of AI's carbon footprint
- Users must validate outputs before using in sensitive or high-stakes applications
---
|