CindyDelage commited on
Commit
51d55d0
·
verified ·
1 Parent(s): 5988328

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -62
README.md CHANGED
@@ -7,65 +7,43 @@ sdk: docker
7
  pinned: false
8
  ---
9
 
10
-
11
- # Random Baseline Model for Climate Disinformation Classification
12
-
13
- ## Model Description
14
-
15
- This is a random baseline model for the Frugal AI Challenge 2024, specifically for the text classification task of identifying climate disinformation. The model serves as a performance floor, randomly assigning labels to text inputs without any learning.
16
-
17
- ### Intended Use
18
-
19
- - **Primary intended uses**: Baseline comparison for climate disinformation classification models
20
- - **Primary intended users**: Researchers and developers participating in the Frugal AI Challenge
21
- - **Out-of-scope use cases**: Not intended for production use or real-world classification tasks
22
-
23
- ## Training Data
24
-
25
- The model uses the QuotaClimat/frugalaichallenge-text-train dataset:
26
- - Size: ~6000 examples
27
- - Split: 80% train, 20% test
28
- - 8 categories of climate disinformation claims
29
-
30
- ### Labels
31
- 0. No relevant claim detected
32
- 1. Global warming is not happening
33
- 2. Not caused by humans
34
- 3. Not bad or beneficial
35
- 4. Solutions harmful/unnecessary
36
- 5. Science is unreliable
37
- 6. Proponents are biased
38
- 7. Fossil fuels are needed
39
-
40
- ## Performance
41
-
42
- ### Metrics
43
- - **Accuracy**: ~12.5% (random chance with 8 classes)
44
- - **Environmental Impact**:
45
- - Emissions tracked in gCO2eq
46
- - Energy consumption tracked in Wh
47
-
48
- ### Model Architecture
49
- The model implements a random choice between the 8 possible labels, serving as the simplest possible baseline.
50
-
51
- ## Environmental Impact
52
-
53
- Environmental impact is tracked using CodeCarbon, measuring:
54
- - Carbon emissions during inference
55
- - Energy consumption during inference
56
-
57
- This tracking helps establish a baseline for the environmental impact of model deployment and inference.
58
-
59
- ## Limitations
60
- - Makes completely random predictions
61
- - No learning or pattern recognition
62
- - No consideration of input text
63
- - Serves only as a baseline reference
64
- - Not suitable for any real-world applications
65
-
66
- ## Ethical Considerations
67
-
68
- - Dataset contains sensitive topics related to climate disinformation
69
- - Model makes random predictions and should not be used for actual classification
70
- - Environmental impact is tracked to promote awareness of AI's carbon footprint
71
- ```
 
7
  pinned: false
8
  ---
9
 
10
+ This model is a fine-tuned version of facebook/wav2vec2-base on an unknown dataset. It achieves the following results on the evaluation set:
11
+
12
+ Loss: 0.3650
13
+ Accuracy: 0.9365
14
+ Model description
15
+ More information needed
16
+
17
+ Intended uses & limitations
18
+ More information needed
19
+
20
+ Training and evaluation data
21
+ More information needed
22
+
23
+ Training procedure
24
+ Training hyperparameters
25
+ The following hyperparameters were used during training:
26
+
27
+ learning_rate: 1e-05
28
+ train_batch_size: 2
29
+ eval_batch_size: 2
30
+ seed: 42
31
+ gradient_accumulation_steps: 4
32
+ total_train_batch_size: 8
33
+ optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
34
+ lr_scheduler_type: linear
35
+ lr_scheduler_warmup_ratio: 0.05
36
+ num_epochs: 7
37
+ Training results
38
+ Training Loss Epoch Step Validation Loss Accuracy
39
+ 1.5669 0.9998 3527 0.2941 0.9211
40
+ 1.3782 1.9998 7054 0.4237 0.9216
41
+ 1.3008 2.9998 10581 0.4684 0.9113
42
+ 1.2606 3.9998 14108 0.4323 0.9204
43
+ 1.0533 4.9998 17635 0.3708 0.9331
44
+ 0.8921 5.9998 21162 0.3650 0.9365
45
+ Framework versions
46
+ Transformers 4.47.1
47
+ Pytorch 2.5.1+cu121
48
+ Datasets 3.2.0
49
+ Tokenizers 0.21.0