AliNajafi commited on
Commit
154bee5
·
1 Parent(s): fa4d2e7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -8
README.md CHANGED
@@ -21,8 +21,6 @@ widget:
21
  ---
22
 
23
 
24
- # TurkishBERTweet
25
-
26
  #### Table of contents
27
  1. [Introduction](#introduction)
28
  2. [Main results](#results)
@@ -36,7 +34,7 @@ widget:
36
  - [HateSpeech Detection](#hs_lora)
37
 
38
  4. [Citation](#citation)
39
- # <a name="introduction"></a> TurkishBERTweet in the shadow of Large Language Models
40
 
41
 
42
  # <a name="results"></a> Main Results
@@ -48,13 +46,13 @@ widget:
48
  # <a name="trainedModels"></a> Model
49
  Model | #params | Arch. | Max length | Pre-training data
50
  ---|---|---|---|---
51
- `VRLLab/TurkishBERTweet` | 163M | base | 128 | 894M Turkish Tweets (uncased)
52
 
53
  # <a name="loraAdapter"></a> Lora Adapters
54
  Model | train f1 | dev f1 | test f1 | Dataset Size
55
  ---|---|---|---|---
56
- `VRLLab/TurkishBERTweet-Lora-SA` | 0.799 | 0.687 | 0.692 | 42,476 Turkish Tweets
57
- `VRLLab/TurkishBERTweet-Lora-HS` | 0.915 | 0.796 | 0.831 | 4,683 Turkish Tweets
58
  # <a name="usage2"></a> Example usage
59
 
60
 
@@ -203,13 +201,14 @@ Yes : kasmayin artik ya kac kere tanik olduk bu azgin tehlikeli “multecilerin
203
  # <a name="citation"></a> Citation
204
  ```bibtex
205
  @article{najafi2022TurkishBERTweet,
206
- title={TurkishBERTweet in the shadow of Large Language Models},
207
  author={Najafi, Ali and Varol, Onur},
208
- journal={arXiv preprint },
209
  year={2023}
210
  }
211
  ```
212
 
 
213
  ## Acknowledgments
214
  We thank [Fatih Amasyali](https://avesis.yildiz.edu.tr/amasyali) for providing access to Tweet Sentiment datasets from Kemik group.
215
  This material is based upon work supported by the Google Cloud Research Credits program with the award GCP19980904. We also thank TUBITAK (121C220 and 222N311) for funding this project.
 
21
  ---
22
 
23
 
 
 
24
  #### Table of contents
25
  1. [Introduction](#introduction)
26
  2. [Main results](#results)
 
34
  - [HateSpeech Detection](#hs_lora)
35
 
36
  4. [Citation](#citation)
37
+ # <a name="introduction"></a> TurkishBERTweet: Fast and Reliable Large Language Model for Social Media Analysis
38
 
39
 
40
  # <a name="results"></a> Main Results
 
46
  # <a name="trainedModels"></a> Model
47
  Model | #params | Arch. | Max length | Pre-training data
48
  ---|---|---|---|---
49
+ [`VRLLab/TurkishBERTweet`](https://huggingface.co/VRLLab/TurkishBERTweet) | 163M | base | 128 | 894M Turkish Tweets (uncased)
50
 
51
  # <a name="loraAdapter"></a> Lora Adapters
52
  Model | train f1 | dev f1 | test f1 | Dataset Size
53
  ---|---|---|---|---
54
+ [`VRLLab/TurkishBERTweet-Lora-SA`](https://huggingface.co/VRLLab/TurkishBERTweet-Lora-SA) | 0.799 | 0.687 | 0.692 | 42,476 Turkish Tweets
55
+ [`VRLLab/TurkishBERTweet-Lora-HS`](https://huggingface.co/VRLLab/TurkishBERTweet-Lora-HS) | 0.915 | 0.796 | 0.831 | 4,683 Turkish Tweets
56
  # <a name="usage2"></a> Example usage
57
 
58
 
 
201
  # <a name="citation"></a> Citation
202
  ```bibtex
203
  @article{najafi2022TurkishBERTweet,
204
+ title={TurkishBERTweet: Fast and Reliable Large Language Model for Social Media Analysis},
205
  author={Najafi, Ali and Varol, Onur},
206
+ journal={arXiv preprint 2311.18063},
207
  year={2023}
208
  }
209
  ```
210
 
211
+
212
  ## Acknowledgments
213
  We thank [Fatih Amasyali](https://avesis.yildiz.edu.tr/amasyali) for providing access to Tweet Sentiment datasets from Kemik group.
214
  This material is based upon work supported by the Google Cloud Research Credits program with the award GCP19980904. We also thank TUBITAK (121C220 and 222N311) for funding this project.