File size: 2,897 Bytes
6fb5d57
a02d840
 
 
 
6fb5d57
a02d840
6fb5d57
 
a02d840
70f5f26
 
 
a02d840
70f5f26
 
 
a02d840
 
 
70f5f26
 
 
 
 
 
 
 
 
a02d840
 
 
 
 
 
 
 
70f5f26
 
 
 
a02d840
 
 
 
70f5f26
 
a02d840
 
 
 
70f5f26
 
 
a02d840
 
 
70f5f26
a02d840
70f5f26
 
a02d840
 
 
70f5f26
 
 
a02d840
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
---
title: Fine-Tuned BERT Model
emoji: 🌍
colorFrom: blue
colorTo: purple
sdk: docker
pinned: true
---

# Fine-Tuned BERT Model for Climate Disinformation Classification

## Model Description

This is a fine-tuned BERT model trained for the Frugal AI Challenge 2024. The model has been fine-tuned on the climate disinformation dataset to classify text inputs into 8 distinct categories related to climate disinformation. It leverages BERT's pretrained language understanding capabilities and has been optimized for accuracy in this domain.

### Intended Use

- **Primary intended uses**: Classifying text inputs to detect specific claims of climate disinformation
- **Primary intended users**: Researchers, developers, and participants in the Frugal AI Challenge
- **Out-of-scope use cases**: Not recommended for tasks outside climate disinformation classification or production-level applications without further evaluation

## Training Data

The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
- Size: ~6000 examples
- Split: 80% train, 20% test
- 8 categories of climate disinformation claims

### Labels
0. No relevant claim detected  
1. Global warming is not happening  
2. Not caused by humans  
3. Not bad or beneficial  
4. Solutions harmful/unnecessary  
5. Science is unreliable  
6. Proponents are biased  
7. Fossil fuels are needed  

## Performance

### Metrics
- **Accuracy**: Achieved XX.X% on the test set (replace `XX.X%` with the actual accuracy from your evaluation)  
- **Environmental Impact**:  
  - Carbon emissions tracked in gCO2eq  
  - Energy consumption tracked in Wh  

### Model Architecture
This model fine-tunes the BERT base architecture (`bert-base-uncased`) for the climate disinformation task. The classifier head includes:  
- Dense layers  
- Dropout for regularization  
- Softmax activation for multi-class classification  

## Environmental Impact

Environmental impact is tracked using CodeCarbon, measuring:  
- Carbon emissions during inference and training  
- Energy consumption during inference and training  

This tracking aligns with the Frugal AI Challenge's commitment to promoting sustainable AI practices.  

## Limitations
- Fine-tuned specifically for climate disinformation; performance on other text classification tasks may degrade  
- Requires computational resources (e.g., GPU) for efficient inference  
- Predictions rely on the training dataset's representativeness; may struggle with unseen or out-of-distribution data  

## Ethical Considerations

- Dataset contains sensitive topics related to climate disinformation  
- Model performance depends on the quality of the dataset and annotation biases  
- Environmental impact during training and inference is disclosed to encourage awareness of AI's carbon footprint  
- Users must validate outputs before using in sensitive or high-stakes applications  

---