File size: 3,504 Bytes
568a8ce
 
01ddcfc
568a8ce
 
 
 
 
82d1ae9
568a8ce
933dfc3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
568a8ce
 
 
 
 
82d1ae9
568a8ce
82d1ae9
568a8ce
82d1ae9
 
568a8ce
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
a952420
 
568a8ce
 
 
 
 
 
 
 
 
82d1ae9
 
 
 
 
 
 
 
 
 
568a8ce
 
 
 
82d1ae9
568a8ce
 
82d1ae9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
---
license: apache-2.0
base_model: distilbert/distilbert-base-uncased
tags:
- generated_from_trainer
metrics:
- accuracy
model-index:
- name: fake-news-detector
  results: []
widget:
- text: >-
    In a shocking turn of events, reports have surfaced suggesting that a
    clandestine meeting of world leaders took place on Mars to discuss plans for
    the colonization of the Red Planet. According to anonymous sources within
    the highest echelons of government, the summit was organized by a coalition
    of space agencies and private corporations aiming to expedite humanity's
    expansion beyond Earth. The meeting purportedly took place in a hidden
    underground facility on Mars, accessible only to a select few individuals
    privy to the ambitious project.
  example_title: Mars Meeting
- text: >-
    In a groundbreaking revelation that has sent shockwaves through the
    scientific community, Dr. Rachel Bennett, a renowned researcher at the
    prestigious Cambridge Institute of Biotechnology, claims to have unlocked
    the elusive secret to eternal youth. According to Dr. Bennett, years of
    tireless research have culminated in the discovery of a revolutionary
    anti-aging compound derived from a rare Amazonian plant known only to
    indigenous tribes. Initial trials on laboratory mice have yielded
    astonishing results, with subjects exhibiting signs of reversed aging and
    enhanced vitality.
  example_title: Dr. Bennett
- text: Apples are orange
  example_title: Oranges are Apples
- text: Donald Trump is the 45th president of the United States.
  example_title: True News
datasets:
- AlexanderHolmes0/true-fake-news
language:
- en
pipeline_tag: text-classification
---

<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->

# fake-news-detector

This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.0027
- Accuracy: 0.9994

## Model description

More information needed

## Intended uses & limitations

More information needed

## Training and evaluation data

More information needed

## Training procedure

### Training hyperparameters

The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 32
- eval_batch_size: 32
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 1

### Training results

| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.201         | 0.09  | 100  | 0.0444          | 0.9901   |
| 0.0319        | 0.19  | 200  | 0.0241          | 0.9938   |
| 0.0222        | 0.28  | 300  | 0.0249          | 0.9932   |
| 0.0094        | 0.38  | 400  | 0.0076          | 0.9984   |
| 0.0042        | 0.47  | 500  | 0.0062          | 0.9988   |
| 0.0076        | 0.57  | 600  | 0.0040          | 0.9988   |
| 0.0095        | 0.66  | 700  | 0.0040          | 0.9990   |
| 0.008         | 0.76  | 800  | 0.0040          | 0.9988   |
| 0.0086        | 0.85  | 900  | 0.0030          | 0.9993   |
| 0.0042        | 0.95  | 1000 | 0.0027          | 0.9994   |


### Framework versions

- Transformers 4.39.1
- Pytorch 2.2.0+cu121
- Datasets 2.18.0
- Tokenizers 0.15.1