MatthiasPicard commited on
Commit
17f802c
·
verified ·
1 Parent(s): 7526ed1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -47
README.md CHANGED
@@ -7,25 +7,14 @@ sdk: docker
7
  pinned: false
8
  ---
9
 
10
-
11
- # Random Baseline Model for Climate Disinformation Classification
12
-
13
  ## Model Description
14
 
15
- This is a random baseline model for the Frugal AI Challenge 2024, specifically for the text classification task of identifying climate disinformation. The model serves as a performance floor, randomly assigning labels to text inputs without any learning.
16
-
17
- ### Intended Use
18
-
19
- - **Primary intended uses**: Baseline comparison for climate disinformation classification models
20
- - **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge
21
- - **Out-of-scope use cases**: Not intended for production use or real-world classification tasks
22
 
23
- ## Training Data
24
 
25
- The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
26
- - Size: ~6000 examples
27
- - Split: 80% train, 20% test
28
- - 8 categories of climate disinformation claims
29
 
30
  ### Labels
31
  0. No relevant claim detected
@@ -36,36 +25,4 @@ The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
36
  5. Science is unreliable
37
  6. Proponents are biased
38
  7. Fossil fuels are needed
39
-
40
- ## Performance
41
-
42
- ### Metrics
43
- - **Accuracy**: ~12.5% (random chance with 8 classes)
44
- - **Environmental Impact**:
45
- - Emissions tracked in gCO2eq
46
- - Energy consumption tracked in Wh
47
-
48
- ### Model Architecture
49
- The model implements a random choice between the 8 possible labels, serving as the simplest possible baseline.
50
-
51
- ## Environmental Impact
52
-
53
- Environmental impact is tracked using CodeCarbon, measuring:
54
- - Carbon emissions during inference
55
- - Energy consumption during inference
56
-
57
- This tracking helps establish a baseline for the environmental impact of model deployment and inference.
58
-
59
- ## Limitations
60
- - Makes completely random predictions
61
- - No learning or pattern recognition
62
- - No consideration of input text
63
- - Serves only as a baseline reference
64
- - Not suitable for any real-world applications
65
-
66
- ## Ethical Considerations
67
-
68
- - Dataset contains sensitive topics related to climate disinformation
69
- - Model makes random predictions and should not be used for actual classification
70
- - Environmental impact is tracked to promote awareness of AI's carbon footprint
71
  ```
 
7
  pinned: false
8
  ---
9
 
 
 
 
10
  ## Model Description
11
 
12
+ This space is dedicated to the text task of the Frugal AI Challenge. The final model employed is a Qwen2.5-3B-Instruct with LoRA adapters trained on a diverse mix of approximately 95,000 samples, encompassing both real and synthetic data.
13
+ The dataset was open-sourced at MatthiasPicard/Frugal-AI-Train-Data-88k. The fine-tuned model, along with training logs, was open-sourced at MatthiasPicard/ModernBERT_frugal_88k.
 
 
 
 
 
14
 
15
+ To optimize inference time, the model was quantized to 8 bits to reduce memory usage and increase performance speed.
16
 
17
+ ### Note: The inference script includes both model and tokenizer loading. As a result, the first evaluation of our model in the submission space will consume more energy than subsequent evaluations.
 
 
 
18
 
19
  ### Labels
20
  0. No relevant claim detected
 
25
  5. Science is unreliable
26
  6. Proponents are biased
27
  7. Fossil fuels are needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
28
  ```