PavanDeepak commited on
Commit
7d53171
·
verified ·
1 Parent(s): f9f78a9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +29 -0
README.md CHANGED
@@ -1,3 +1,32 @@
1
  ---
2
  license: mit
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
  ---
4
+ BERT-based Text Classification Model
5
+ This model is a fine-tuned version of the bert-base-uncased model, specifically adapted for text classification across a diverse set of categories. The model has been trained on a rich dataset collected from multiple sources, including the News Category Dataset on Kaggle and various other websites.
6
+
7
+ The model classifies text into one of the following 12 categories:
8
+
9
+ Food
10
+ Videogames & Shows
11
+ Kids and fun
12
+ Homestyle
13
+ Travel
14
+ Health
15
+ Charity
16
+ Electronics & Technology
17
+ Sports
18
+ Cultural & Music
19
+ Education
20
+ Convenience
21
+ The model has demonstrated robust performance with an accuracy of 0.721459, F1 score of 0.659451, precision of 0.707620, and recall of 0.635155.
22
+
23
+ Model Architecture
24
+ The model leverages the BertForSequenceClassification architecture, It has been fine-tuned on the aforementioned dataset, with the following key configuration parameters:
25
+
26
+ Hidden size: 768
27
+ Number of attention heads: 12
28
+ Number of hidden layers: 12
29
+ Max position embeddings: 512
30
+ Type vocab size: 2
31
+ Vocab size: 30522
32
+ The model uses the GELU activation function in its hidden layers and applies dropout with a probability of 0.1 to the attention probabilities to prevent overfitting.