Spaces:
Sleeping
Sleeping
Update README.md
Browse files
README.md
CHANGED
@@ -7,25 +7,14 @@ sdk: docker
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
10 |
-
|
11 |
-
# Random Baseline Model for Climate Disinformation Classification
|
12 |
-
|
13 |
## Model Description
|
14 |
|
15 |
-
This is
|
16 |
-
|
17 |
-
### Intended Use
|
18 |
-
|
19 |
-
- **Primary intended uses**: Baseline comparison for climate disinformation classification models
|
20 |
-
- **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge
|
21 |
-
- **Out-of-scope use cases**: Not intended for production use or real-world classification tasks
|
22 |
|
23 |
-
|
24 |
|
25 |
-
The model
|
26 |
-
- Size: ~6000 examples
|
27 |
-
- Split: 80% train, 20% test
|
28 |
-
- 8 categories of climate disinformation claims
|
29 |
|
30 |
### Labels
|
31 |
0. No relevant claim detected
|
@@ -36,36 +25,4 @@ The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
|
|
36 |
5. Science is unreliable
|
37 |
6. Proponents are biased
|
38 |
7. Fossil fuels are needed
|
39 |
-
|
40 |
-
## Performance
|
41 |
-
|
42 |
-
### Metrics
|
43 |
-
- **Accuracy**: ~12.5% (random chance with 8 classes)
|
44 |
-
- **Environmental Impact**:
|
45 |
-
- Emissions tracked in gCO2eq
|
46 |
-
- Energy consumption tracked in Wh
|
47 |
-
|
48 |
-
### Model Architecture
|
49 |
-
The model implements a random choice between the 8 possible labels, serving as the simplest possible baseline.
|
50 |
-
|
51 |
-
## Environmental Impact
|
52 |
-
|
53 |
-
Environmental impact is tracked using CodeCarbon, measuring:
|
54 |
-
- Carbon emissions during inference
|
55 |
-
- Energy consumption during inference
|
56 |
-
|
57 |
-
This tracking helps establish a baseline for the environmental impact of model deployment and inference.
|
58 |
-
|
59 |
-
## Limitations
|
60 |
-
- Makes completely random predictions
|
61 |
-
- No learning or pattern recognition
|
62 |
-
- No consideration of input text
|
63 |
-
- Serves only as a baseline reference
|
64 |
-
- Not suitable for any real-world applications
|
65 |
-
|
66 |
-
## Ethical Considerations
|
67 |
-
|
68 |
-
- Dataset contains sensitive topics related to climate disinformation
|
69 |
-
- Model makes random predictions and should not be used for actual classification
|
70 |
-
- Environmental impact is tracked to promote awareness of AI's carbon footprint
|
71 |
```
|
|
|
7 |
pinned: false
|
8 |
---
|
9 |
|
|
|
|
|
|
|
10 |
## Model Description
|
11 |
|
12 |
+
This space is dedicated to the text task of the Frugal AI Challenge. The final model employed is a Qwen2.5-3B-Instruct with LoRA adapters trained on a diverse mix of approximately 95,000 samples, encompassing both real and synthetic data.
|
13 |
+
The dataset was open-sourced at MatthiasPicard/Frugal-AI-Train-Data-88k. The fine-tuned model, along with training logs, was open-sourced at MatthiasPicard/ModernBERT_frugal_88k.
|
|
|
|
|
|
|
|
|
|
|
14 |
|
15 |
+
To optimize inference time, the model was quantized to 8 bits to reduce memory usage and increase performance speed.
|
16 |
|
17 |
+
### Note: The inference script includes both model and tokenizer loading. As a result, the first evaluation of our model in the submission space will consume more energy than subsequent evaluations.
|
|
|
|
|
|
|
18 |
|
19 |
### Labels
|
20 |
0. No relevant claim detected
|
|
|
25 |
5. Science is unreliable
|
26 |
6. Proponents are biased
|
27 |
7. Fossil fuels are needed
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
28 |
```
|