asahi417 commited on
Commit
b615d10
·
1 Parent(s): f8d386e
Files changed (46) hide show
  1. .gitattributes +43 -0
  2. data/tweet_emoji/test.jsonl +3 -0
  3. data/tweet_emoji/test_1.jsonl +3 -0
  4. data/tweet_emoji/test_2.jsonl +3 -0
  5. data/tweet_emoji/test_3.jsonl +3 -0
  6. data/tweet_emoji/test_4.jsonl +3 -0
  7. data/tweet_emoji/train.jsonl +3 -0
  8. data/tweet_emoji/validation.jsonl +3 -0
  9. data/tweet_emoji_test0_seed0/test.jsonl +3 -0
  10. data/tweet_emoji_test0_seed0/train.jsonl +3 -0
  11. data/tweet_emoji_test0_seed0/validation.jsonl +3 -0
  12. data/tweet_emoji_test0_seed1/test.jsonl +3 -0
  13. data/tweet_emoji_test0_seed1/train.jsonl +3 -0
  14. data/tweet_emoji_test0_seed1/validation.jsonl +3 -0
  15. data/tweet_emoji_test0_seed2/test.jsonl +3 -0
  16. data/tweet_emoji_test0_seed2/train.jsonl +3 -0
  17. data/tweet_emoji_test0_seed2/validation.jsonl +3 -0
  18. data/tweet_emoji_test1_seed0/test.jsonl +3 -0
  19. data/tweet_emoji_test1_seed0/train.jsonl +3 -0
  20. data/tweet_emoji_test1_seed0/validation.jsonl +3 -0
  21. data/tweet_emoji_test1_seed1/test.jsonl +3 -0
  22. data/tweet_emoji_test1_seed1/train.jsonl +3 -0
  23. data/tweet_emoji_test1_seed1/validation.jsonl +3 -0
  24. data/tweet_emoji_test1_seed2/test.jsonl +3 -0
  25. data/tweet_emoji_test1_seed2/train.jsonl +3 -0
  26. data/tweet_emoji_test1_seed2/validation.jsonl +3 -0
  27. data/tweet_emoji_test2_seed0/test.jsonl +3 -0
  28. data/tweet_emoji_test2_seed0/train.jsonl +3 -0
  29. data/tweet_emoji_test2_seed0/validation.jsonl +3 -0
  30. data/tweet_emoji_test2_seed1/test.jsonl +3 -0
  31. data/tweet_emoji_test2_seed1/train.jsonl +3 -0
  32. data/tweet_emoji_test2_seed1/validation.jsonl +3 -0
  33. data/tweet_emoji_test2_seed2/test.jsonl +3 -0
  34. data/tweet_emoji_test2_seed2/train.jsonl +3 -0
  35. data/tweet_emoji_test2_seed2/validation.jsonl +3 -0
  36. data/tweet_emoji_test3_seed0/test.jsonl +3 -0
  37. data/tweet_emoji_test3_seed0/train.jsonl +3 -0
  38. data/tweet_emoji_test3_seed0/validation.jsonl +3 -0
  39. data/tweet_emoji_test3_seed1/test.jsonl +3 -0
  40. data/tweet_emoji_test3_seed1/train.jsonl +3 -0
  41. data/tweet_emoji_test3_seed1/validation.jsonl +3 -0
  42. data/tweet_emoji_test3_seed2/test.jsonl +3 -0
  43. data/tweet_emoji_test3_seed2/train.jsonl +3 -0
  44. data/tweet_emoji_test3_seed2/validation.jsonl +3 -0
  45. process/tweet_emoji.py +62 -44
  46. tweet_temporal_shift.py +30 -2
.gitattributes CHANGED
@@ -558,3 +558,46 @@ data/tweet_hate_test1_seed2/train.jsonl filter=lfs diff=lfs merge=lfs -text
558
  data/tweet_hate_test1_seed2/validation.jsonl filter=lfs diff=lfs merge=lfs -text
559
  data/tweet_hate_test2_seed2/test.jsonl filter=lfs diff=lfs merge=lfs -text
560
  data/tweet_hate_test3_seed1/train.jsonl filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
558
  data/tweet_hate_test1_seed2/validation.jsonl filter=lfs diff=lfs merge=lfs -text
559
  data/tweet_hate_test2_seed2/test.jsonl filter=lfs diff=lfs merge=lfs -text
560
  data/tweet_hate_test3_seed1/train.jsonl filter=lfs diff=lfs merge=lfs -text
561
+ data/tweet_emoji_test1_seed1/test.jsonl filter=lfs diff=lfs merge=lfs -text
562
+ data/tweet_emoji_test3_seed2/train.jsonl filter=lfs diff=lfs merge=lfs -text
563
+ data/tweet_emoji_test2_seed1/train.jsonl filter=lfs diff=lfs merge=lfs -text
564
+ data/tweet_emoji_test3_seed1/train.jsonl filter=lfs diff=lfs merge=lfs -text
565
+ data/tweet_emoji/test_2.jsonl filter=lfs diff=lfs merge=lfs -text
566
+ data/tweet_emoji/test_4.jsonl filter=lfs diff=lfs merge=lfs -text
567
+ data/tweet_emoji_test0_seed0/train.jsonl filter=lfs diff=lfs merge=lfs -text
568
+ data/tweet_emoji_test0_seed2/validation.jsonl filter=lfs diff=lfs merge=lfs -text
569
+ data/tweet_emoji_test2_seed0/train.jsonl filter=lfs diff=lfs merge=lfs -text
570
+ data/tweet_emoji_test2_seed2/train.jsonl filter=lfs diff=lfs merge=lfs -text
571
+ data/tweet_emoji_test0_seed1/test.jsonl filter=lfs diff=lfs merge=lfs -text
572
+ data/tweet_emoji_test0_seed1/train.jsonl filter=lfs diff=lfs merge=lfs -text
573
+ data/tweet_emoji_test0_seed2/test.jsonl filter=lfs diff=lfs merge=lfs -text
574
+ data/tweet_emoji_test1_seed2/validation.jsonl filter=lfs diff=lfs merge=lfs -text
575
+ data/tweet_emoji/train.jsonl filter=lfs diff=lfs merge=lfs -text
576
+ data/tweet_emoji_test1_seed0/test.jsonl filter=lfs diff=lfs merge=lfs -text
577
+ data/tweet_emoji_test2_seed0/test.jsonl filter=lfs diff=lfs merge=lfs -text
578
+ data/tweet_emoji_test3_seed0/train.jsonl filter=lfs diff=lfs merge=lfs -text
579
+ data/tweet_emoji_test3_seed1/validation.jsonl filter=lfs diff=lfs merge=lfs -text
580
+ data/tweet_emoji_test3_seed2/test.jsonl filter=lfs diff=lfs merge=lfs -text
581
+ data/tweet_emoji/test_3.jsonl filter=lfs diff=lfs merge=lfs -text
582
+ data/tweet_emoji_test1_seed0/train.jsonl filter=lfs diff=lfs merge=lfs -text
583
+ data/tweet_emoji_test1_seed2/test.jsonl filter=lfs diff=lfs merge=lfs -text
584
+ data/tweet_emoji_test3_seed0/test.jsonl filter=lfs diff=lfs merge=lfs -text
585
+ data/tweet_emoji_test3_seed2/validation.jsonl filter=lfs diff=lfs merge=lfs -text
586
+ data/tweet_emoji_test0_seed0/test.jsonl filter=lfs diff=lfs merge=lfs -text
587
+ data/tweet_emoji_test2_seed0/validation.jsonl filter=lfs diff=lfs merge=lfs -text
588
+ data/tweet_emoji_test2_seed1/test.jsonl filter=lfs diff=lfs merge=lfs -text
589
+ data/tweet_emoji_test2_seed1/validation.jsonl filter=lfs diff=lfs merge=lfs -text
590
+ data/tweet_emoji_test3_seed0/validation.jsonl filter=lfs diff=lfs merge=lfs -text
591
+ data/tweet_emoji/test_1.jsonl filter=lfs diff=lfs merge=lfs -text
592
+ data/tweet_emoji_test0_seed2/train.jsonl filter=lfs diff=lfs merge=lfs -text
593
+ data/tweet_emoji_test1_seed0/validation.jsonl filter=lfs diff=lfs merge=lfs -text
594
+ data/tweet_emoji_test1_seed1/validation.jsonl filter=lfs diff=lfs merge=lfs -text
595
+ data/tweet_emoji_test1_seed1/train.jsonl filter=lfs diff=lfs merge=lfs -text
596
+ data/tweet_emoji_test1_seed2/train.jsonl filter=lfs diff=lfs merge=lfs -text
597
+ data/tweet_emoji_test2_seed2/test.jsonl filter=lfs diff=lfs merge=lfs -text
598
+ data/tweet_emoji_test2_seed2/validation.jsonl filter=lfs diff=lfs merge=lfs -text
599
+ data/tweet_emoji/test.jsonl filter=lfs diff=lfs merge=lfs -text
600
+ data/tweet_emoji/validation.jsonl filter=lfs diff=lfs merge=lfs -text
601
+ data/tweet_emoji_test0_seed0/validation.jsonl filter=lfs diff=lfs merge=lfs -text
602
+ data/tweet_emoji_test0_seed1/validation.jsonl filter=lfs diff=lfs merge=lfs -text
603
+ data/tweet_emoji_test3_seed1/test.jsonl filter=lfs diff=lfs merge=lfs -text
data/tweet_emoji/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75d2c96a4cc54548f994bb1f8e8db8873c2ccbe64030522885e3fe897184126e
3
+ size 7239725
data/tweet_emoji/test_1.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe7fe4990c50474dff72cb933eb936dbd341a828c1c10a06462d1a54a1b21e3f
3
+ size 1803375
data/tweet_emoji/test_2.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea84eb6cb539da4d35d8c3417e5e66abf261a7088b19b3c77d9d4b4d319b1aba
3
+ size 1811980
data/tweet_emoji/test_3.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b054c53207be5df6842f5f096e36e11faccc0b490730f229d9b47ce15fd87663
3
+ size 1814319
data/tweet_emoji/test_4.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fc7579556ff64ebe67f997f3bba6b209d3e881f3de5f5c4abb78a24ca270dff
3
+ size 1810048
data/tweet_emoji/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:87a40027d9aaa0e3bacd8f323543d5db99c07209b58a41d450be4c24c8fd4b8e
3
+ size 5750763
data/tweet_emoji/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39b0e8967e102fb8a1fb94d98afaddf8808d3f92cb21ece1c668d83303ee86ee
3
+ size 1438080
data/tweet_emoji_test0_seed0/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe7fe4990c50474dff72cb933eb936dbd341a828c1c10a06462d1a54a1b21e3f
3
+ size 1803375
data/tweet_emoji_test0_seed0/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aedece58ad60f4b71e408abb6254b8f9322cfa1e96c0235df364f0cac1f0252e
3
+ size 5782726
data/tweet_emoji_test0_seed0/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b88bd5f4cd7275321b5f68dd4db37502bb5c27781cfc7afb11abe0a9f2471ac
3
+ size 1414688
data/tweet_emoji_test0_seed1/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe7fe4990c50474dff72cb933eb936dbd341a828c1c10a06462d1a54a1b21e3f
3
+ size 1803375
data/tweet_emoji_test0_seed1/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fad46cbc7af7648af516e0fc30b3c89d28fe99ee0b06da8e49a4b4f3ef59c238
3
+ size 5768033
data/tweet_emoji_test0_seed1/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4f1876a8fde3e6f97df2c0541da65c61091e63345a73e21cdc3fbacbf7ff41c
3
+ size 1405398
data/tweet_emoji_test0_seed2/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe7fe4990c50474dff72cb933eb936dbd341a828c1c10a06462d1a54a1b21e3f
3
+ size 1803375
data/tweet_emoji_test0_seed2/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07bb580977903ab54f4c74e2497d544f650e6cfdff6892b31c4381f2f63a9c1c
3
+ size 5762067
data/tweet_emoji_test0_seed2/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d5771425ec9c1cf471b7fd9df667e940c8d2f5d0979a2f355a90748ef69b20b
3
+ size 1412828
data/tweet_emoji_test1_seed0/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea84eb6cb539da4d35d8c3417e5e66abf261a7088b19b3c77d9d4b4d319b1aba
3
+ size 1811980
data/tweet_emoji_test1_seed0/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:409e48a47610e5a02abbd881e15e18bd7c845d8a3e70818d95c1022feb9a3712
3
+ size 5762805
data/tweet_emoji_test1_seed0/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:13e6aa4a2c9b67f230e9f635154a54cfeceaa32828242594572162cd0e69e3e7
3
+ size 1410268
data/tweet_emoji_test1_seed1/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea84eb6cb539da4d35d8c3417e5e66abf261a7088b19b3c77d9d4b4d319b1aba
3
+ size 1811980
data/tweet_emoji_test1_seed1/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:764e4f705b4a8986f9fe2b4da3dc3f86ef9355d18255cfdc0ef265a22936a058
3
+ size 5784682
data/tweet_emoji_test1_seed1/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b34e2987661b5487579d1ab916ea714510b43bd57b56e2338884347b22155241
3
+ size 1406216
data/tweet_emoji_test1_seed2/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea84eb6cb539da4d35d8c3417e5e66abf261a7088b19b3c77d9d4b4d319b1aba
3
+ size 1811980
data/tweet_emoji_test1_seed2/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3464fe224b36e5172d20d9c729da4f9ba1960dfe4c931d010c28a1f65aadf2ae
3
+ size 5758640
data/tweet_emoji_test1_seed2/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe5f9c2997d1d9055bd6da95620715c6b9842f2c2954e351b83e8a19d42d5307
3
+ size 1408730
data/tweet_emoji_test2_seed0/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b054c53207be5df6842f5f096e36e11faccc0b490730f229d9b47ce15fd87663
3
+ size 1814319
data/tweet_emoji_test2_seed0/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55e101b368193b6dc5e6415c96303732b9388670c50f462cdc9462f8481e39bc
3
+ size 5772421
data/tweet_emoji_test2_seed0/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed8888e6dcd88aa81dead857db387d791d1bb0d4c7639fb01d6fd525394ad3bb
3
+ size 1411077
data/tweet_emoji_test2_seed1/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b054c53207be5df6842f5f096e36e11faccc0b490730f229d9b47ce15fd87663
3
+ size 1814319
data/tweet_emoji_test2_seed1/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e6b63b7d88a631ef76f694ffd0a46fa63029661e04593cd853fd51eb146af06f
3
+ size 5778443
data/tweet_emoji_test2_seed1/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9eb8474fc70cc118f960d245718a9274ffb175af88d45c465971b00034577e8
3
+ size 1406950
data/tweet_emoji_test2_seed2/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b054c53207be5df6842f5f096e36e11faccc0b490730f229d9b47ce15fd87663
3
+ size 1814319
data/tweet_emoji_test2_seed2/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34c717be198dfb803616e92b98b734240dd4e35e9010330fcdaad3672fcd11a3
3
+ size 5779840
data/tweet_emoji_test2_seed2/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3edc65d2f4af319fa5084ce53a606954b0c27447fbf337d5b87e9b5fa728567
3
+ size 1404787
data/tweet_emoji_test3_seed0/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fc7579556ff64ebe67f997f3bba6b209d3e881f3de5f5c4abb78a24ca270dff
3
+ size 1810048
data/tweet_emoji_test3_seed0/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56fe7dd45e8de4bfed78e58ab4f5b5911c6da42ee0939ffd0709bc9b31cecb1b
3
+ size 5777027
data/tweet_emoji_test3_seed0/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67043409b923f87f2fef25261991c2c0e0350da52765d9cd5e517ef87187abdb
3
+ size 1405620
data/tweet_emoji_test3_seed1/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fc7579556ff64ebe67f997f3bba6b209d3e881f3de5f5c4abb78a24ca270dff
3
+ size 1810048
data/tweet_emoji_test3_seed1/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1fa9a36b436b69190f425fc7f8cac3c9dc607439c6ecc94cde07c5c507accf3f
3
+ size 5781243
data/tweet_emoji_test3_seed1/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:60ac675d741ad1018c41dd096be805f10539757b1d4ce8e77aabdd4932bfc7c9
3
+ size 1411991
data/tweet_emoji_test3_seed2/test.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2fc7579556ff64ebe67f997f3bba6b209d3e881f3de5f5c4abb78a24ca270dff
3
+ size 1810048
data/tweet_emoji_test3_seed2/train.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:891bb59c8405ac2c82951107f72a193f640b7bc9ab3840856306bfd4cdc2ffe8
3
+ size 5780044
data/tweet_emoji_test3_seed2/validation.jsonl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5416d48f42e9eafaa583a1ee2ca2c2a49457010f4aa3f3e6d2bd6a48bfc859e3
3
+ size 1405958
process/tweet_emoji.py CHANGED
@@ -1,58 +1,70 @@
1
- # TODO
2
  import json
3
  import os
4
  from random import shuffle, seed
5
 
 
 
6
 
7
- with open("data/tweet_sentiment/test.jsonl") as f:
8
- test = [json.loads(i) for i in f if len(i)]
9
- with open("data/tweet_sentiment/test_1.jsonl") as f:
10
- test_1 = [json.loads(i) for i in f if len(i)]
11
- with open("data/tweet_sentiment/test_2.jsonl") as f:
12
- test_2 = [json.loads(i) for i in f if len(i)]
13
- with open("data/tweet_sentiment/test_3.jsonl") as f:
14
- test_3 = [json.loads(i) for i in f if len(i)]
15
- with open("data/tweet_sentiment/test_4.jsonl") as f:
16
- test_4 = [json.loads(i) for i in f if len(i)]
17
- with open("data/tweet_sentiment/train.jsonl") as f:
18
- train = [json.loads(i) for i in f if len(i)]
19
- with open("data/tweet_sentiment/validation.jsonl") as f:
20
- validation = [json.loads(i) for i in f if len(i)]
 
 
 
 
 
21
 
 
 
 
 
22
 
23
- os.makedirs(f"data/tweet_sentiment_small", exist_ok=True)
24
- with open(f"data/tweet_sentiment_small/test.jsonl", "w") as f:
25
- f.write("\n".join([json.dumps(i) for i in test]))
 
 
 
 
 
 
 
26
 
27
- with open(f"data/tweet_sentiment_small/test_1.jsonl", "w") as f:
 
 
 
 
 
 
 
 
 
 
28
  f.write("\n".join([json.dumps(i) for i in test_1]))
29
- with open(f"data/tweet_sentiment_small/test_2.jsonl", "w") as f:
30
  f.write("\n".join([json.dumps(i) for i in test_2]))
31
- with open(f"data/tweet_sentiment_small/test_3.jsonl", "w") as f:
32
  f.write("\n".join([json.dumps(i) for i in test_3]))
33
- with open(f"data/tweet_sentiment_small/test_4.jsonl", "w") as f:
34
  f.write("\n".join([json.dumps(i) for i in test_4]))
35
-
36
- with open(f"data/tweet_sentiment_small/validation.jsonl", "w") as f:
37
- f.write("\n".join([json.dumps(i) for i in validation]))
38
-
39
- # down sample training set
40
- n_train_p = 2500
41
- n_train_n = 2500
42
- seed(123)
43
- shuffle(train)
44
- train_p = [i for i in train if i["gold_label_binary"] == 0][:n_train_p]
45
- train_n = [i for i in train if i["gold_label_binary"] == 1][:n_train_n]
46
- train = train_p + train_n
47
- shuffle(train)
48
- with open(f"data/tweet_sentiment_small/train.jsonl", "w") as f:
49
  f.write("\n".join([json.dumps(i) for i in train]))
 
 
50
 
51
 
52
- n_train = len(train)
53
- n_validation = len(validation)
54
- n_test = int(len(test)/4)
55
-
56
  def sampler(dataset_test, r_seed):
57
  seed(r_seed)
58
  shuffle(dataset_test)
@@ -71,11 +83,17 @@ for n, _test in enumerate([
71
  test_1 + test_2 + test_4,
72
  test_1 + test_2 + test_3]):
73
  for s in range(3):
74
- os.makedirs(f"data/tweet_sentiment_small_test{n}_seed{s}", exist_ok=True)
75
  _train, _valid = sampler(_test, s)
76
- with open(f"data/tweet_sentiment_small_test{n}_seed{s}/train.jsonl", "w") as f:
77
  f.write("\n".join([json.dumps(i) for i in _train]))
78
- with open(f"data/tweet_sentiment_small_test{n}_seed{s}/validation.jsonl", "w") as f:
79
  f.write("\n".join([json.dumps(i) for i in _valid]))
80
- with open(f"data/tweet_sentiment_small_test{n}_seed{s}/test.jsonl", "w") as f:
81
  f.write("\n".join([json.dumps(i) for i in id2test[n]]))
 
 
 
 
 
 
 
 
1
  import json
2
  import os
3
  from random import shuffle, seed
4
 
5
+ import pandas as pd
6
+ from datasets import load_dataset
7
 
8
+ test = load_dataset("cardiffnlp/super_tweeteval", "tweet_emoji", split="test").shuffle(seed=42)
9
+ test = list(test.to_pandas().T.to_dict().values())
10
+ train = load_dataset("cardiffnlp/super_tweeteval", "tweet_emoji", split="train").shuffle(seed=42)
11
+ train = list(train.to_pandas().T.to_dict().values())
12
+ validation = load_dataset("cardiffnlp/super_tweeteval", "tweet_emoji", split="validation").shuffle(seed=42)
13
+ validation = list(validation.to_pandas().T.to_dict().values())
14
+ full = train + test + validation
15
+ df = pd.DataFrame(full)
16
+ df["date_dt"] = pd.to_datetime(df.date)
17
+ df = df.sort_values(by="date_dt")
18
+ dist_date = df.groupby("date_dt").size()
19
+ total_n = len(df)
20
+ n = 0
21
+ while True:
22
+ n += 1
23
+ if dist_date[:n].sum() > total_n/2:
24
+ break
25
+ split_date = dist_date.index[n]
26
+ print(split_date)
27
 
28
+ train = df[df["date_dt"] <= split_date]
29
+ test = df[df["date_dt"] > split_date]
30
+ print(train.date_dt.min(), train.date_dt.max())
31
+ print(test.date_dt.min(), test.date_dt.max())
32
 
33
+ train.pop("date_dt")
34
+ test.pop("date_dt")
35
+ train = list(train.T.to_dict().values())
36
+ test = list(test.T.to_dict().values())
37
+
38
+ seed(42)
39
+ shuffle(train)
40
+ shuffle(test)
41
+ valid = train[:int(len(train)*0.2)]
42
+ train = train[len(valid):]
43
 
44
+ n_test = int(len(test)/4)
45
+ n_train = len(train)
46
+ n_validation = len(valid)
47
+ test_1 = test[:n_test]
48
+ test_2 = test[n_test:n_test*2]
49
+ test_3 = test[n_test*2:n_test*3]
50
+ test_4 = test[n_test*3:]
51
+ os.makedirs("data/tweet_emoji", exist_ok=True)
52
+ with open("data/tweet_emoji/test.jsonl", "w") as f:
53
+ f.write("\n".join([json.dumps(i) for i in test]))
54
+ with open("data/tweet_emoji/test_1.jsonl", "w") as f:
55
  f.write("\n".join([json.dumps(i) for i in test_1]))
56
+ with open("data/tweet_emoji/test_2.jsonl", "w") as f:
57
  f.write("\n".join([json.dumps(i) for i in test_2]))
58
+ with open("data/tweet_emoji/test_3.jsonl", "w") as f:
59
  f.write("\n".join([json.dumps(i) for i in test_3]))
60
+ with open("data/tweet_emoji/test_4.jsonl", "w") as f:
61
  f.write("\n".join([json.dumps(i) for i in test_4]))
62
+ with open("data/tweet_emoji/train.jsonl", "w") as f:
 
 
 
 
 
 
 
 
 
 
 
 
 
63
  f.write("\n".join([json.dumps(i) for i in train]))
64
+ with open("data/tweet_emoji/validation.jsonl", "w") as f:
65
+ f.write("\n".join([json.dumps(i) for i in valid]))
66
 
67
 
 
 
 
 
68
  def sampler(dataset_test, r_seed):
69
  seed(r_seed)
70
  shuffle(dataset_test)
 
83
  test_1 + test_2 + test_4,
84
  test_1 + test_2 + test_3]):
85
  for s in range(3):
86
+ os.makedirs(f"data/tweet_emoji_test{n}_seed{s}", exist_ok=True)
87
  _train, _valid = sampler(_test, s)
88
+ with open(f"data/tweet_emoji_test{n}_seed{s}/train.jsonl", "w") as f:
89
  f.write("\n".join([json.dumps(i) for i in _train]))
90
+ with open(f"data/tweet_emoji_test{n}_seed{s}/validation.jsonl", "w") as f:
91
  f.write("\n".join([json.dumps(i) for i in _valid]))
92
+ with open(f"data/tweet_emoji_test{n}_seed{s}/test.jsonl", "w") as f:
93
  f.write("\n".join([json.dumps(i) for i in id2test[n]]))
94
+
95
+
96
+
97
+
98
+
99
+
tweet_temporal_shift.py CHANGED
@@ -2,7 +2,7 @@
2
  import json
3
  import datasets
4
 
5
- _VERSION = "1.0.8"
6
  _TWEET_TEMPORAL_DESCRIPTION = """"""
7
  _TWEET_TEMPORAL_CITATION = """"""
8
  _TWEET_TOPIC_DESCRIPTION = """
@@ -69,6 +69,10 @@ _TWEET_HATE_DESCRIPTION = """TBA"""
69
  _TWEET_HATE_CITATION = """\
70
  TBA
71
  """
 
 
 
 
72
  _ROOT_URL = "https://huggingface.co/datasets/tweettemposhift/tweet_temporal_shift/resolve/main/data"
73
 
74
 
@@ -142,6 +146,13 @@ class TweetTemporalShift(datasets.GeneratorBasedBuilder):
142
  features=["gold_label_binary", "text", "date"],
143
  data_url=f"{_ROOT_URL}/tweet_sentiment_small",
144
  ),
 
 
 
 
 
 
 
145
  ]
146
  for s in range(3):
147
  for i in range(4):
@@ -188,6 +199,13 @@ class TweetTemporalShift(datasets.GeneratorBasedBuilder):
188
  features=["gold_label_binary", "text", "date"],
189
  data_url=f"{_ROOT_URL}/tweet_sentiment_small_test{i}_seed{s}",
190
  ),
 
 
 
 
 
 
 
191
  ]
192
 
193
  def _info(self):
@@ -197,7 +215,8 @@ class TweetTemporalShift(datasets.GeneratorBasedBuilder):
197
  "arts_&_culture", "business_&_entrepreneurs", "celebrity_&_pop_culture", "diaries_&_daily_life",
198
  "family", "fashion_&_style", "film_tv_&_video", "fitness_&_health", "food_&_dining", "gaming",
199
  "learning_&_educational", "music", "news_&_social_concern", "other_hobbies", "relationships",
200
- "science_&_technology", "sports", "travel_&_adventure", "youth_&_student_life"]
 
201
  features["gold_label_list"] = datasets.Sequence(
202
  datasets.features.ClassLabel(names=names))
203
  elif "hate" in self.config.name:
@@ -206,6 +225,15 @@ class TweetTemporalShift(datasets.GeneratorBasedBuilder):
206
  features["date"] = datasets.Value("string")
207
  features["id"] = datasets.Value("string")
208
  features["source"] = datasets.Value("int32")
 
 
 
 
 
 
 
 
 
209
  elif "sentiment" in self.config.name:
210
  features["text"] = datasets.Value("string")
211
  features["gold_label_binary"] = datasets.Value("int32")
 
2
  import json
3
  import datasets
4
 
5
+ _VERSION = "1.0.9"
6
  _TWEET_TEMPORAL_DESCRIPTION = """"""
7
  _TWEET_TEMPORAL_CITATION = """"""
8
  _TWEET_TOPIC_DESCRIPTION = """
 
69
  _TWEET_HATE_CITATION = """\
70
  TBA
71
  """
72
+ _TWEET_EMOJI_DESCRIPTION = """TBA"""
73
+ _TWEET_EMOJI_CITATION = """\
74
+ TBA
75
+ """
76
  _ROOT_URL = "https://huggingface.co/datasets/tweettemposhift/tweet_temporal_shift/resolve/main/data"
77
 
78
 
 
146
  features=["gold_label_binary", "text", "date"],
147
  data_url=f"{_ROOT_URL}/tweet_sentiment_small",
148
  ),
149
+ TweetTemporalShiftConfig(
150
+ name="emoji_temporal",
151
+ description=_TWEET_EMOJI_DESCRIPTION,
152
+ citation=_TWEET_EMOJI_CITATION,
153
+ features=["text", "date", "gold_label"],
154
+ data_url=f"{_ROOT_URL}/tweet_emoji",
155
+ ),
156
  ]
157
  for s in range(3):
158
  for i in range(4):
 
199
  features=["gold_label_binary", "text", "date"],
200
  data_url=f"{_ROOT_URL}/tweet_sentiment_small_test{i}_seed{s}",
201
  ),
202
+ TweetTemporalShiftConfig(
203
+ name=f"emoji_temporal_random{i}_seed{s}",
204
+ description=_TWEET_EMOJI_DESCRIPTION,
205
+ citation=_TWEET_EMOJI_CITATION,
206
+ features=["text", "date", "gold_label"],
207
+ data_url=f"{_ROOT_URL}/tweet_emoji_test{i}_seed{s}",
208
+ ),
209
  ]
210
 
211
  def _info(self):
 
215
  "arts_&_culture", "business_&_entrepreneurs", "celebrity_&_pop_culture", "diaries_&_daily_life",
216
  "family", "fashion_&_style", "film_tv_&_video", "fitness_&_health", "food_&_dining", "gaming",
217
  "learning_&_educational", "music", "news_&_social_concern", "other_hobbies", "relationships",
218
+ "science_&_technology", "sports", "travel_&_adventure", "youth_&_student_life"
219
+ ]
220
  features["gold_label_list"] = datasets.Sequence(
221
  datasets.features.ClassLabel(names=names))
222
  elif "hate" in self.config.name:
 
225
  features["date"] = datasets.Value("string")
226
  features["id"] = datasets.Value("string")
227
  features["source"] = datasets.Value("int32")
228
+ elif "emoji" in self.config.name:
229
+ features["text"] = datasets.Value("string")
230
+ features["date"] = datasets.Value("string")
231
+ url_map = "https://huggingface.co/datasets/cardiffnlp/super_tweet_eval/resolve/main/data/tweet_emoji/map.txt"
232
+ dl_manager = datasets.utils.download_manager.DownloadManager()
233
+ with open(dl_manager.download(url_map)) as f:
234
+ label_classes = f.readlines()
235
+ label_classes = [x.strip('\n') for x in label_classes]
236
+ features['gold_label'] = datasets.features.ClassLabel(names=label_classes)
237
  elif "sentiment" in self.config.name:
238
  features["text"] = datasets.Value("string")
239
  features["gold_label_binary"] = datasets.Value("int32")