Jingmei commited on
Commit
3832801
1 Parent(s): bda2c85

Training in progress, step 10

Browse files
Files changed (36) hide show
  1. adapter_config.json +29 -0
  2. adapter_model.safetensors +3 -0
  3. trainer_peft.log +217 -0
  4. training_args.bin +3 -0
  5. transformed_cache/test_chunked_00000_of_00008.arrow +3 -0
  6. transformed_cache/test_chunked_00001_of_00008.arrow +3 -0
  7. transformed_cache/test_chunked_00002_of_00008.arrow +3 -0
  8. transformed_cache/test_chunked_00003_of_00008.arrow +3 -0
  9. transformed_cache/test_chunked_00004_of_00008.arrow +3 -0
  10. transformed_cache/test_chunked_00005_of_00008.arrow +3 -0
  11. transformed_cache/test_chunked_00006_of_00008.arrow +3 -0
  12. transformed_cache/test_chunked_00007_of_00008.arrow +3 -0
  13. transformed_cache/test_tokenized_00000_of_00008.arrow +3 -0
  14. transformed_cache/test_tokenized_00001_of_00008.arrow +3 -0
  15. transformed_cache/test_tokenized_00002_of_00008.arrow +3 -0
  16. transformed_cache/test_tokenized_00003_of_00008.arrow +3 -0
  17. transformed_cache/test_tokenized_00004_of_00008.arrow +3 -0
  18. transformed_cache/test_tokenized_00005_of_00008.arrow +3 -0
  19. transformed_cache/test_tokenized_00006_of_00008.arrow +3 -0
  20. transformed_cache/test_tokenized_00007_of_00008.arrow +3 -0
  21. transformed_cache/train_chunked_00000_of_00008.arrow +3 -0
  22. transformed_cache/train_chunked_00001_of_00008.arrow +3 -0
  23. transformed_cache/train_chunked_00002_of_00008.arrow +3 -0
  24. transformed_cache/train_chunked_00003_of_00008.arrow +3 -0
  25. transformed_cache/train_chunked_00004_of_00008.arrow +3 -0
  26. transformed_cache/train_chunked_00005_of_00008.arrow +3 -0
  27. transformed_cache/train_chunked_00006_of_00008.arrow +3 -0
  28. transformed_cache/train_chunked_00007_of_00008.arrow +3 -0
  29. transformed_cache/train_tokenized_00000_of_00008.arrow +3 -0
  30. transformed_cache/train_tokenized_00001_of_00008.arrow +3 -0
  31. transformed_cache/train_tokenized_00002_of_00008.arrow +3 -0
  32. transformed_cache/train_tokenized_00003_of_00008.arrow +3 -0
  33. transformed_cache/train_tokenized_00004_of_00008.arrow +3 -0
  34. transformed_cache/train_tokenized_00005_of_00008.arrow +3 -0
  35. transformed_cache/train_tokenized_00006_of_00008.arrow +3 -0
  36. transformed_cache/train_tokenized_00007_of_00008.arrow +3 -0
adapter_config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "chaoyi-wu/PMC_LLAMA_7B",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 32,
14
+ "lora_dropout": 0.1,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 8,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "v_proj",
24
+ "q_proj"
25
+ ],
26
+ "task_type": "CAUSAL_LM",
27
+ "use_dora": false,
28
+ "use_rslora": false
29
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad97e300b7291fc510547c93dfef18124d35201660770dbc90813de392da8736
3
+ size 16794200
trainer_peft.log ADDED
@@ -0,0 +1,217 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2024-05-29 18:45 - Cuda check
2
+ 2024-05-29 18:45 - True
3
+ 2024-05-29 18:45 - 1
4
+ 2024-05-29 18:45 - Configue Model and tokenizer
5
+ 2024-05-29 18:45 - Memory usage in 0.00 GB
6
+ 2024-05-29 18:45 - Dataset loaded successfully:
7
+ train-Jingmei/Pandemic_Wiki
8
+ test -Jingmei/Pandemic_WHO
9
+ 2024-05-29 18:46 - Tokenize data: DatasetDict({
10
+ train: Dataset({
11
+ features: ['input_ids', 'attention_mask'],
12
+ num_rows: 2152
13
+ })
14
+ test: Dataset({
15
+ features: ['input_ids', 'attention_mask'],
16
+ num_rows: 8264
17
+ })
18
+ })
19
+ 2024-05-29 18:49 - Split data into chunks:DatasetDict({
20
+ train: Dataset({
21
+ features: ['input_ids', 'attention_mask'],
22
+ num_rows: 24863
23
+ })
24
+ test: Dataset({
25
+ features: ['input_ids', 'attention_mask'],
26
+ num_rows: 198964
27
+ })
28
+ })
29
+ 2024-05-29 18:49 - Setup PEFT
30
+ 2024-05-29 18:49 - Setup optimizer
31
+ 2024-05-29 18:49 - Start training
32
+ 2024-05-29 18:57 - Cuda check
33
+ 2024-05-29 18:57 - True
34
+ 2024-05-29 18:57 - 1
35
+ 2024-05-29 18:57 - Configue Model and tokenizer
36
+ 2024-05-29 18:57 - Memory usage in 25.17 GB
37
+ 2024-05-29 18:57 - Dataset loaded successfully:
38
+ train-Jingmei/Pandemic_Wiki
39
+ test -Jingmei/Pandemic_WHO
40
+ 2024-05-29 18:57 - Tokenize data: DatasetDict({
41
+ train: Dataset({
42
+ features: ['input_ids', 'attention_mask'],
43
+ num_rows: 2152
44
+ })
45
+ test: Dataset({
46
+ features: ['input_ids', 'attention_mask'],
47
+ num_rows: 8264
48
+ })
49
+ })
50
+ 2024-05-29 18:57 - Split data into chunks:DatasetDict({
51
+ train: Dataset({
52
+ features: ['input_ids', 'attention_mask'],
53
+ num_rows: 24863
54
+ })
55
+ test: Dataset({
56
+ features: ['input_ids', 'attention_mask'],
57
+ num_rows: 198964
58
+ })
59
+ })
60
+ 2024-05-29 18:57 - Setup PEFT
61
+ 2024-05-29 18:57 - Setup optimizer
62
+ 2024-05-29 18:57 - Start training
63
+ 2024-05-29 19:04 - Cuda check
64
+ 2024-05-29 19:04 - True
65
+ 2024-05-29 19:04 - 1
66
+ 2024-05-29 19:04 - Configue Model and tokenizer
67
+ 2024-05-29 19:04 - Memory usage in 25.17 GB
68
+ 2024-05-29 19:04 - Dataset loaded successfully:
69
+ train-Jingmei/Pandemic_Wiki
70
+ test -Jingmei/Pandemic_WHO
71
+ 2024-05-29 19:04 - Tokenize data: DatasetDict({
72
+ train: Dataset({
73
+ features: ['input_ids', 'attention_mask'],
74
+ num_rows: 2152
75
+ })
76
+ test: Dataset({
77
+ features: ['input_ids', 'attention_mask'],
78
+ num_rows: 8264
79
+ })
80
+ })
81
+ 2024-05-29 19:04 - Split data into chunks:DatasetDict({
82
+ train: Dataset({
83
+ features: ['input_ids', 'attention_mask'],
84
+ num_rows: 24863
85
+ })
86
+ test: Dataset({
87
+ features: ['input_ids', 'attention_mask'],
88
+ num_rows: 198964
89
+ })
90
+ })
91
+ 2024-05-29 19:04 - Setup PEFT
92
+ 2024-05-29 19:04 - Setup optimizer
93
+ 2024-05-29 19:04 - Start training
94
+ 2024-05-29 19:10 - Cuda check
95
+ 2024-05-29 19:10 - True
96
+ 2024-05-29 19:10 - 1
97
+ 2024-05-29 19:10 - Configue Model and tokenizer
98
+ 2024-05-29 19:10 - Memory usage in 25.17 GB
99
+ 2024-05-29 19:10 - Dataset loaded successfully:
100
+ train-Jingmei/Pandemic_Wiki
101
+ test -Jingmei/Pandemic_WHO
102
+ 2024-05-29 19:10 - Tokenize data: DatasetDict({
103
+ train: Dataset({
104
+ features: ['input_ids', 'attention_mask'],
105
+ num_rows: 2152
106
+ })
107
+ test: Dataset({
108
+ features: ['input_ids', 'attention_mask'],
109
+ num_rows: 8264
110
+ })
111
+ })
112
+ 2024-05-29 19:10 - Split data into chunks:DatasetDict({
113
+ train: Dataset({
114
+ features: ['input_ids', 'attention_mask'],
115
+ num_rows: 24863
116
+ })
117
+ test: Dataset({
118
+ features: ['input_ids', 'attention_mask'],
119
+ num_rows: 198964
120
+ })
121
+ })
122
+ 2024-05-29 19:10 - Setup PEFT
123
+ 2024-05-29 19:10 - Setup optimizer
124
+ 2024-05-29 19:10 - Start training
125
+ 2024-05-29 19:16 - Cuda check
126
+ 2024-05-29 19:16 - True
127
+ 2024-05-29 19:16 - 1
128
+ 2024-05-29 19:16 - Configue Model and tokenizer
129
+ 2024-05-29 19:16 - Memory usage in 25.17 GB
130
+ 2024-05-29 19:16 - Dataset loaded successfully:
131
+ train-Jingmei/Pandemic_Wiki
132
+ test -Jingmei/Pandemic_WHO
133
+ 2024-05-29 19:16 - Tokenize data: DatasetDict({
134
+ train: Dataset({
135
+ features: ['input_ids', 'attention_mask'],
136
+ num_rows: 2152
137
+ })
138
+ test: Dataset({
139
+ features: ['input_ids', 'attention_mask'],
140
+ num_rows: 8264
141
+ })
142
+ })
143
+ 2024-05-29 19:16 - Split data into chunks:DatasetDict({
144
+ train: Dataset({
145
+ features: ['input_ids', 'attention_mask'],
146
+ num_rows: 24863
147
+ })
148
+ test: Dataset({
149
+ features: ['input_ids', 'attention_mask'],
150
+ num_rows: 198964
151
+ })
152
+ })
153
+ 2024-05-29 19:16 - Setup PEFT
154
+ 2024-05-29 19:16 - Setup optimizer
155
+ 2024-05-29 19:16 - Start training
156
+ 2024-05-29 19:22 - Cuda check
157
+ 2024-05-29 19:22 - True
158
+ 2024-05-29 19:22 - 1
159
+ 2024-05-29 19:22 - Configue Model and tokenizer
160
+ 2024-05-29 19:22 - Memory usage in 25.17 GB
161
+ 2024-05-29 19:22 - Dataset loaded successfully:
162
+ train-Jingmei/Pandemic_Wiki
163
+ test -Jingmei/Pandemic_WHO
164
+ 2024-05-29 19:22 - Tokenize data: DatasetDict({
165
+ train: Dataset({
166
+ features: ['input_ids', 'attention_mask'],
167
+ num_rows: 2152
168
+ })
169
+ test: Dataset({
170
+ features: ['input_ids', 'attention_mask'],
171
+ num_rows: 8264
172
+ })
173
+ })
174
+ 2024-05-29 19:22 - Split data into chunks:DatasetDict({
175
+ train: Dataset({
176
+ features: ['input_ids', 'attention_mask'],
177
+ num_rows: 24863
178
+ })
179
+ test: Dataset({
180
+ features: ['input_ids', 'attention_mask'],
181
+ num_rows: 198964
182
+ })
183
+ })
184
+ 2024-05-29 19:22 - Setup PEFT
185
+ 2024-05-29 19:22 - Setup optimizer
186
+ 2024-05-29 19:22 - Start training
187
+ 2024-05-29 19:29 - Cuda check
188
+ 2024-05-29 19:29 - True
189
+ 2024-05-29 19:29 - 1
190
+ 2024-05-29 19:29 - Configue Model and tokenizer
191
+ 2024-05-29 19:29 - Memory usage in 25.17 GB
192
+ 2024-05-29 19:29 - Dataset loaded successfully:
193
+ train-Jingmei/Pandemic_Wiki
194
+ test -Jingmei/Pandemic_WHO
195
+ 2024-05-29 19:29 - Tokenize data: DatasetDict({
196
+ train: Dataset({
197
+ features: ['input_ids', 'attention_mask'],
198
+ num_rows: 2152
199
+ })
200
+ test: Dataset({
201
+ features: ['input_ids', 'attention_mask'],
202
+ num_rows: 8264
203
+ })
204
+ })
205
+ 2024-05-29 19:29 - Split data into chunks:DatasetDict({
206
+ train: Dataset({
207
+ features: ['input_ids', 'attention_mask'],
208
+ num_rows: 24863
209
+ })
210
+ test: Dataset({
211
+ features: ['input_ids', 'attention_mask'],
212
+ num_rows: 198964
213
+ })
214
+ })
215
+ 2024-05-29 19:29 - Setup PEFT
216
+ 2024-05-29 19:29 - Setup optimizer
217
+ 2024-05-29 19:29 - Start training
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d57b2583d86e2ed6d84da0f503e8cc8b5ab9a9038e8eae344efc26164f159e6
3
+ size 5176
transformed_cache/test_chunked_00000_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dae2e9968e4f98dd4f16f89da1efc7aa266ab4d0d26b7f30b51e309c6f828c27
3
+ size 62148672
transformed_cache/test_chunked_00001_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:11db4c2496f57ecbc395ef6fcf09d9e00a87a226a605445f53a43b3e85022dc8
3
+ size 59367216
transformed_cache/test_chunked_00002_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b40e232ff015f3c9bd7b9a1e1e3027bd690b95110d8e7602d6b8b4a181e3b3bc
3
+ size 68269120
transformed_cache/test_chunked_00003_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3ade749ffe68b29e54cc4c5076e3d14f88a6ce323b5c578d71b5395017aa532
3
+ size 80368176
transformed_cache/test_chunked_00004_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b933533de1fa4a3ab31ecd0a428a9bc09a5bc5a543c3f609300df5294af90db7
3
+ size 55943760
transformed_cache/test_chunked_00005_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b1794b5460b5b248bbe30abdcbcd5a9473c1f0301faa1c1f58dda59e64c4ad5
3
+ size 62541568
transformed_cache/test_chunked_00006_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:485d66f28cad901af230c63ce95c958968af004a3c647855c513f6a9ecd4cb98
3
+ size 58971744
transformed_cache/test_chunked_00007_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6eb3d0575e477eb043fe5cfedd358c449c7b758a5299fade16672fe3402979fa
3
+ size 63397008
transformed_cache/test_tokenized_00000_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f19e4e96ea46d5b1cb50a3d5f173d13cc65d5105bf59bc4287252fa1b6dd05e
3
+ size 61957552
transformed_cache/test_tokenized_00001_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4de30f634c4db0778fb02c6647561302c0e6d82d9edffb28241e2abc3ef74f87
3
+ size 59187104
transformed_cache/test_tokenized_00002_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a96c44a43cf8d01a7f5567064c039c413d77b368e2b6ae86d1e440cdd5053650
3
+ size 68057936
transformed_cache/test_tokenized_00003_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f4daeadd552b3364911e35740457a497ee896c85c721da89603f7b0103af0ee
3
+ size 80121008
transformed_cache/test_tokenized_00004_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04f9aa02a689092fb09c4cce9d4560c0e9ac3560d7f5b2e3827bfd690a2689c7
3
+ size 55773176
transformed_cache/test_tokenized_00005_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eeb72ac2d0ecb57de088b4f6558020a5946e833d28dfeabebad937c3c553190
3
+ size 62349352
transformed_cache/test_tokenized_00006_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1175f22330a67782d122ddd79d40c6bac37cf81a99012f452b3549055eeecff
3
+ size 58791688
transformed_cache/test_tokenized_00007_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb2c05949063ade79e6e710254bda5068062d53d3862bb7a094eee3162833500
3
+ size 63205208
transformed_cache/train_chunked_00000_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b95e6793aeafcaa2d4964868dc52637199ce2c6a47272fcec347c47d57eda89
3
+ size 7184208
transformed_cache/train_chunked_00001_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2236d0790dfa94c4a687688ee379b2b34cc5b34e668809c6b658073909079b38
3
+ size 8260512
transformed_cache/train_chunked_00002_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4471933f4dde7bdc2581ec4c18993149fd5020c2cc5b68db1ed54def6a73c925
3
+ size 7898416
transformed_cache/train_chunked_00003_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81165c3d8c0ef0373f4cc00e15e3b38bf10bb4c203f31c26f83a6d7abb56dcd8
3
+ size 8237392
transformed_cache/train_chunked_00004_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65b919ee6dae99de6a36facd60126885a306d893e64da7408ca7ead656f4ce3e
3
+ size 8342688
transformed_cache/train_chunked_00005_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48b3ccabb6a5f30597e582f07e48480cb5210c3c483d230ebfca37d6c1fbd072
3
+ size 8848576
transformed_cache/train_chunked_00006_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37de1908b28108464569ab1b34f8f8d8f3ddc1c604282f971acef60fb6d98621
3
+ size 7538592
transformed_cache/train_chunked_00007_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:403da53b850cd65cb3b6f1e9c41c35c44c0c9da4c8c7c84a7c96cfd0adb48648
3
+ size 7551440
transformed_cache/train_tokenized_00000_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab46d354dc75989c187c176b231b957d40cf5fb6f1cbb7c900970ea9e3355e3b
3
+ size 7164512
transformed_cache/train_tokenized_00001_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94da0daa98038663cb57dfe1608c8df4f923c2d4db39a030349bc7a25374be06
3
+ size 8236248
transformed_cache/train_tokenized_00002_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26f696795dee6c02fe594b04ac52f23c4d3b0d4f7daa66f42b9798a2abb1cf93
3
+ size 7876672
transformed_cache/train_tokenized_00003_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d477a0a9b8ff7cf1da7281b895b11e05affe217d81f44fcfc68eab70ecf0b3e
3
+ size 8214000
transformed_cache/train_tokenized_00004_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b30406de2a4fb0ae56a08aca3a7457d6d63aa0ac66a7b950c84cd4252de96ed
3
+ size 8318008
transformed_cache/train_tokenized_00005_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19c3e9c2abf180bda95e87458b6f610b45173b3eb4d7c10e37db10ee1b07cf64
3
+ size 8823040
transformed_cache/train_tokenized_00006_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a714384172fed68834fc866a5a799ffe9cb301598b41d5433d24a74727fa15a4
3
+ size 7516936
transformed_cache/train_tokenized_00007_of_00008.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a15c328603fef8a816eae22386f22485b37040adc7237397620e129656c9a565
3
+ size 7529536