Training in progress, step 10
Browse files- adapter_config.json +29 -0
- adapter_model.safetensors +3 -0
- trainer_peft.log +217 -0
- training_args.bin +3 -0
- transformed_cache/test_chunked_00000_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00001_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00002_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00003_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00004_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00005_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00006_of_00008.arrow +3 -0
- transformed_cache/test_chunked_00007_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00000_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00001_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00002_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00003_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00004_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00005_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00006_of_00008.arrow +3 -0
- transformed_cache/test_tokenized_00007_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00000_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00001_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00002_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00003_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00004_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00005_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00006_of_00008.arrow +3 -0
- transformed_cache/train_chunked_00007_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00000_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00001_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00002_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00003_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00004_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00005_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00006_of_00008.arrow +3 -0
- transformed_cache/train_tokenized_00007_of_00008.arrow +3 -0
adapter_config.json
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"alpha_pattern": {},
|
3 |
+
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "chaoyi-wu/PMC_LLAMA_7B",
|
5 |
+
"bias": "none",
|
6 |
+
"fan_in_fan_out": false,
|
7 |
+
"inference_mode": true,
|
8 |
+
"init_lora_weights": true,
|
9 |
+
"layer_replication": null,
|
10 |
+
"layers_pattern": null,
|
11 |
+
"layers_to_transform": null,
|
12 |
+
"loftq_config": {},
|
13 |
+
"lora_alpha": 32,
|
14 |
+
"lora_dropout": 0.1,
|
15 |
+
"megatron_config": null,
|
16 |
+
"megatron_core": "megatron.core",
|
17 |
+
"modules_to_save": null,
|
18 |
+
"peft_type": "LORA",
|
19 |
+
"r": 8,
|
20 |
+
"rank_pattern": {},
|
21 |
+
"revision": null,
|
22 |
+
"target_modules": [
|
23 |
+
"v_proj",
|
24 |
+
"q_proj"
|
25 |
+
],
|
26 |
+
"task_type": "CAUSAL_LM",
|
27 |
+
"use_dora": false,
|
28 |
+
"use_rslora": false
|
29 |
+
}
|
adapter_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ad97e300b7291fc510547c93dfef18124d35201660770dbc90813de392da8736
|
3 |
+
size 16794200
|
trainer_peft.log
ADDED
@@ -0,0 +1,217 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
2024-05-29 18:45 - Cuda check
|
2 |
+
2024-05-29 18:45 - True
|
3 |
+
2024-05-29 18:45 - 1
|
4 |
+
2024-05-29 18:45 - Configue Model and tokenizer
|
5 |
+
2024-05-29 18:45 - Memory usage in 0.00 GB
|
6 |
+
2024-05-29 18:45 - Dataset loaded successfully:
|
7 |
+
train-Jingmei/Pandemic_Wiki
|
8 |
+
test -Jingmei/Pandemic_WHO
|
9 |
+
2024-05-29 18:46 - Tokenize data: DatasetDict({
|
10 |
+
train: Dataset({
|
11 |
+
features: ['input_ids', 'attention_mask'],
|
12 |
+
num_rows: 2152
|
13 |
+
})
|
14 |
+
test: Dataset({
|
15 |
+
features: ['input_ids', 'attention_mask'],
|
16 |
+
num_rows: 8264
|
17 |
+
})
|
18 |
+
})
|
19 |
+
2024-05-29 18:49 - Split data into chunks:DatasetDict({
|
20 |
+
train: Dataset({
|
21 |
+
features: ['input_ids', 'attention_mask'],
|
22 |
+
num_rows: 24863
|
23 |
+
})
|
24 |
+
test: Dataset({
|
25 |
+
features: ['input_ids', 'attention_mask'],
|
26 |
+
num_rows: 198964
|
27 |
+
})
|
28 |
+
})
|
29 |
+
2024-05-29 18:49 - Setup PEFT
|
30 |
+
2024-05-29 18:49 - Setup optimizer
|
31 |
+
2024-05-29 18:49 - Start training
|
32 |
+
2024-05-29 18:57 - Cuda check
|
33 |
+
2024-05-29 18:57 - True
|
34 |
+
2024-05-29 18:57 - 1
|
35 |
+
2024-05-29 18:57 - Configue Model and tokenizer
|
36 |
+
2024-05-29 18:57 - Memory usage in 25.17 GB
|
37 |
+
2024-05-29 18:57 - Dataset loaded successfully:
|
38 |
+
train-Jingmei/Pandemic_Wiki
|
39 |
+
test -Jingmei/Pandemic_WHO
|
40 |
+
2024-05-29 18:57 - Tokenize data: DatasetDict({
|
41 |
+
train: Dataset({
|
42 |
+
features: ['input_ids', 'attention_mask'],
|
43 |
+
num_rows: 2152
|
44 |
+
})
|
45 |
+
test: Dataset({
|
46 |
+
features: ['input_ids', 'attention_mask'],
|
47 |
+
num_rows: 8264
|
48 |
+
})
|
49 |
+
})
|
50 |
+
2024-05-29 18:57 - Split data into chunks:DatasetDict({
|
51 |
+
train: Dataset({
|
52 |
+
features: ['input_ids', 'attention_mask'],
|
53 |
+
num_rows: 24863
|
54 |
+
})
|
55 |
+
test: Dataset({
|
56 |
+
features: ['input_ids', 'attention_mask'],
|
57 |
+
num_rows: 198964
|
58 |
+
})
|
59 |
+
})
|
60 |
+
2024-05-29 18:57 - Setup PEFT
|
61 |
+
2024-05-29 18:57 - Setup optimizer
|
62 |
+
2024-05-29 18:57 - Start training
|
63 |
+
2024-05-29 19:04 - Cuda check
|
64 |
+
2024-05-29 19:04 - True
|
65 |
+
2024-05-29 19:04 - 1
|
66 |
+
2024-05-29 19:04 - Configue Model and tokenizer
|
67 |
+
2024-05-29 19:04 - Memory usage in 25.17 GB
|
68 |
+
2024-05-29 19:04 - Dataset loaded successfully:
|
69 |
+
train-Jingmei/Pandemic_Wiki
|
70 |
+
test -Jingmei/Pandemic_WHO
|
71 |
+
2024-05-29 19:04 - Tokenize data: DatasetDict({
|
72 |
+
train: Dataset({
|
73 |
+
features: ['input_ids', 'attention_mask'],
|
74 |
+
num_rows: 2152
|
75 |
+
})
|
76 |
+
test: Dataset({
|
77 |
+
features: ['input_ids', 'attention_mask'],
|
78 |
+
num_rows: 8264
|
79 |
+
})
|
80 |
+
})
|
81 |
+
2024-05-29 19:04 - Split data into chunks:DatasetDict({
|
82 |
+
train: Dataset({
|
83 |
+
features: ['input_ids', 'attention_mask'],
|
84 |
+
num_rows: 24863
|
85 |
+
})
|
86 |
+
test: Dataset({
|
87 |
+
features: ['input_ids', 'attention_mask'],
|
88 |
+
num_rows: 198964
|
89 |
+
})
|
90 |
+
})
|
91 |
+
2024-05-29 19:04 - Setup PEFT
|
92 |
+
2024-05-29 19:04 - Setup optimizer
|
93 |
+
2024-05-29 19:04 - Start training
|
94 |
+
2024-05-29 19:10 - Cuda check
|
95 |
+
2024-05-29 19:10 - True
|
96 |
+
2024-05-29 19:10 - 1
|
97 |
+
2024-05-29 19:10 - Configue Model and tokenizer
|
98 |
+
2024-05-29 19:10 - Memory usage in 25.17 GB
|
99 |
+
2024-05-29 19:10 - Dataset loaded successfully:
|
100 |
+
train-Jingmei/Pandemic_Wiki
|
101 |
+
test -Jingmei/Pandemic_WHO
|
102 |
+
2024-05-29 19:10 - Tokenize data: DatasetDict({
|
103 |
+
train: Dataset({
|
104 |
+
features: ['input_ids', 'attention_mask'],
|
105 |
+
num_rows: 2152
|
106 |
+
})
|
107 |
+
test: Dataset({
|
108 |
+
features: ['input_ids', 'attention_mask'],
|
109 |
+
num_rows: 8264
|
110 |
+
})
|
111 |
+
})
|
112 |
+
2024-05-29 19:10 - Split data into chunks:DatasetDict({
|
113 |
+
train: Dataset({
|
114 |
+
features: ['input_ids', 'attention_mask'],
|
115 |
+
num_rows: 24863
|
116 |
+
})
|
117 |
+
test: Dataset({
|
118 |
+
features: ['input_ids', 'attention_mask'],
|
119 |
+
num_rows: 198964
|
120 |
+
})
|
121 |
+
})
|
122 |
+
2024-05-29 19:10 - Setup PEFT
|
123 |
+
2024-05-29 19:10 - Setup optimizer
|
124 |
+
2024-05-29 19:10 - Start training
|
125 |
+
2024-05-29 19:16 - Cuda check
|
126 |
+
2024-05-29 19:16 - True
|
127 |
+
2024-05-29 19:16 - 1
|
128 |
+
2024-05-29 19:16 - Configue Model and tokenizer
|
129 |
+
2024-05-29 19:16 - Memory usage in 25.17 GB
|
130 |
+
2024-05-29 19:16 - Dataset loaded successfully:
|
131 |
+
train-Jingmei/Pandemic_Wiki
|
132 |
+
test -Jingmei/Pandemic_WHO
|
133 |
+
2024-05-29 19:16 - Tokenize data: DatasetDict({
|
134 |
+
train: Dataset({
|
135 |
+
features: ['input_ids', 'attention_mask'],
|
136 |
+
num_rows: 2152
|
137 |
+
})
|
138 |
+
test: Dataset({
|
139 |
+
features: ['input_ids', 'attention_mask'],
|
140 |
+
num_rows: 8264
|
141 |
+
})
|
142 |
+
})
|
143 |
+
2024-05-29 19:16 - Split data into chunks:DatasetDict({
|
144 |
+
train: Dataset({
|
145 |
+
features: ['input_ids', 'attention_mask'],
|
146 |
+
num_rows: 24863
|
147 |
+
})
|
148 |
+
test: Dataset({
|
149 |
+
features: ['input_ids', 'attention_mask'],
|
150 |
+
num_rows: 198964
|
151 |
+
})
|
152 |
+
})
|
153 |
+
2024-05-29 19:16 - Setup PEFT
|
154 |
+
2024-05-29 19:16 - Setup optimizer
|
155 |
+
2024-05-29 19:16 - Start training
|
156 |
+
2024-05-29 19:22 - Cuda check
|
157 |
+
2024-05-29 19:22 - True
|
158 |
+
2024-05-29 19:22 - 1
|
159 |
+
2024-05-29 19:22 - Configue Model and tokenizer
|
160 |
+
2024-05-29 19:22 - Memory usage in 25.17 GB
|
161 |
+
2024-05-29 19:22 - Dataset loaded successfully:
|
162 |
+
train-Jingmei/Pandemic_Wiki
|
163 |
+
test -Jingmei/Pandemic_WHO
|
164 |
+
2024-05-29 19:22 - Tokenize data: DatasetDict({
|
165 |
+
train: Dataset({
|
166 |
+
features: ['input_ids', 'attention_mask'],
|
167 |
+
num_rows: 2152
|
168 |
+
})
|
169 |
+
test: Dataset({
|
170 |
+
features: ['input_ids', 'attention_mask'],
|
171 |
+
num_rows: 8264
|
172 |
+
})
|
173 |
+
})
|
174 |
+
2024-05-29 19:22 - Split data into chunks:DatasetDict({
|
175 |
+
train: Dataset({
|
176 |
+
features: ['input_ids', 'attention_mask'],
|
177 |
+
num_rows: 24863
|
178 |
+
})
|
179 |
+
test: Dataset({
|
180 |
+
features: ['input_ids', 'attention_mask'],
|
181 |
+
num_rows: 198964
|
182 |
+
})
|
183 |
+
})
|
184 |
+
2024-05-29 19:22 - Setup PEFT
|
185 |
+
2024-05-29 19:22 - Setup optimizer
|
186 |
+
2024-05-29 19:22 - Start training
|
187 |
+
2024-05-29 19:29 - Cuda check
|
188 |
+
2024-05-29 19:29 - True
|
189 |
+
2024-05-29 19:29 - 1
|
190 |
+
2024-05-29 19:29 - Configue Model and tokenizer
|
191 |
+
2024-05-29 19:29 - Memory usage in 25.17 GB
|
192 |
+
2024-05-29 19:29 - Dataset loaded successfully:
|
193 |
+
train-Jingmei/Pandemic_Wiki
|
194 |
+
test -Jingmei/Pandemic_WHO
|
195 |
+
2024-05-29 19:29 - Tokenize data: DatasetDict({
|
196 |
+
train: Dataset({
|
197 |
+
features: ['input_ids', 'attention_mask'],
|
198 |
+
num_rows: 2152
|
199 |
+
})
|
200 |
+
test: Dataset({
|
201 |
+
features: ['input_ids', 'attention_mask'],
|
202 |
+
num_rows: 8264
|
203 |
+
})
|
204 |
+
})
|
205 |
+
2024-05-29 19:29 - Split data into chunks:DatasetDict({
|
206 |
+
train: Dataset({
|
207 |
+
features: ['input_ids', 'attention_mask'],
|
208 |
+
num_rows: 24863
|
209 |
+
})
|
210 |
+
test: Dataset({
|
211 |
+
features: ['input_ids', 'attention_mask'],
|
212 |
+
num_rows: 198964
|
213 |
+
})
|
214 |
+
})
|
215 |
+
2024-05-29 19:29 - Setup PEFT
|
216 |
+
2024-05-29 19:29 - Setup optimizer
|
217 |
+
2024-05-29 19:29 - Start training
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6d57b2583d86e2ed6d84da0f503e8cc8b5ab9a9038e8eae344efc26164f159e6
|
3 |
+
size 5176
|
transformed_cache/test_chunked_00000_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dae2e9968e4f98dd4f16f89da1efc7aa266ab4d0d26b7f30b51e309c6f828c27
|
3 |
+
size 62148672
|
transformed_cache/test_chunked_00001_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11db4c2496f57ecbc395ef6fcf09d9e00a87a226a605445f53a43b3e85022dc8
|
3 |
+
size 59367216
|
transformed_cache/test_chunked_00002_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b40e232ff015f3c9bd7b9a1e1e3027bd690b95110d8e7602d6b8b4a181e3b3bc
|
3 |
+
size 68269120
|
transformed_cache/test_chunked_00003_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d3ade749ffe68b29e54cc4c5076e3d14f88a6ce323b5c578d71b5395017aa532
|
3 |
+
size 80368176
|
transformed_cache/test_chunked_00004_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b933533de1fa4a3ab31ecd0a428a9bc09a5bc5a543c3f609300df5294af90db7
|
3 |
+
size 55943760
|
transformed_cache/test_chunked_00005_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b1794b5460b5b248bbe30abdcbcd5a9473c1f0301faa1c1f58dda59e64c4ad5
|
3 |
+
size 62541568
|
transformed_cache/test_chunked_00006_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:485d66f28cad901af230c63ce95c958968af004a3c647855c513f6a9ecd4cb98
|
3 |
+
size 58971744
|
transformed_cache/test_chunked_00007_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6eb3d0575e477eb043fe5cfedd358c449c7b758a5299fade16672fe3402979fa
|
3 |
+
size 63397008
|
transformed_cache/test_tokenized_00000_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f19e4e96ea46d5b1cb50a3d5f173d13cc65d5105bf59bc4287252fa1b6dd05e
|
3 |
+
size 61957552
|
transformed_cache/test_tokenized_00001_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4de30f634c4db0778fb02c6647561302c0e6d82d9edffb28241e2abc3ef74f87
|
3 |
+
size 59187104
|
transformed_cache/test_tokenized_00002_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a96c44a43cf8d01a7f5567064c039c413d77b368e2b6ae86d1e440cdd5053650
|
3 |
+
size 68057936
|
transformed_cache/test_tokenized_00003_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4f4daeadd552b3364911e35740457a497ee896c85c721da89603f7b0103af0ee
|
3 |
+
size 80121008
|
transformed_cache/test_tokenized_00004_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:04f9aa02a689092fb09c4cce9d4560c0e9ac3560d7f5b2e3827bfd690a2689c7
|
3 |
+
size 55773176
|
transformed_cache/test_tokenized_00005_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0eeb72ac2d0ecb57de088b4f6558020a5946e833d28dfeabebad937c3c553190
|
3 |
+
size 62349352
|
transformed_cache/test_tokenized_00006_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1175f22330a67782d122ddd79d40c6bac37cf81a99012f452b3549055eeecff
|
3 |
+
size 58791688
|
transformed_cache/test_tokenized_00007_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fb2c05949063ade79e6e710254bda5068062d53d3862bb7a094eee3162833500
|
3 |
+
size 63205208
|
transformed_cache/train_chunked_00000_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3b95e6793aeafcaa2d4964868dc52637199ce2c6a47272fcec347c47d57eda89
|
3 |
+
size 7184208
|
transformed_cache/train_chunked_00001_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2236d0790dfa94c4a687688ee379b2b34cc5b34e668809c6b658073909079b38
|
3 |
+
size 8260512
|
transformed_cache/train_chunked_00002_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4471933f4dde7bdc2581ec4c18993149fd5020c2cc5b68db1ed54def6a73c925
|
3 |
+
size 7898416
|
transformed_cache/train_chunked_00003_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:81165c3d8c0ef0373f4cc00e15e3b38bf10bb4c203f31c26f83a6d7abb56dcd8
|
3 |
+
size 8237392
|
transformed_cache/train_chunked_00004_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65b919ee6dae99de6a36facd60126885a306d893e64da7408ca7ead656f4ce3e
|
3 |
+
size 8342688
|
transformed_cache/train_chunked_00005_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:48b3ccabb6a5f30597e582f07e48480cb5210c3c483d230ebfca37d6c1fbd072
|
3 |
+
size 8848576
|
transformed_cache/train_chunked_00006_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:37de1908b28108464569ab1b34f8f8d8f3ddc1c604282f971acef60fb6d98621
|
3 |
+
size 7538592
|
transformed_cache/train_chunked_00007_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:403da53b850cd65cb3b6f1e9c41c35c44c0c9da4c8c7c84a7c96cfd0adb48648
|
3 |
+
size 7551440
|
transformed_cache/train_tokenized_00000_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab46d354dc75989c187c176b231b957d40cf5fb6f1cbb7c900970ea9e3355e3b
|
3 |
+
size 7164512
|
transformed_cache/train_tokenized_00001_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:94da0daa98038663cb57dfe1608c8df4f923c2d4db39a030349bc7a25374be06
|
3 |
+
size 8236248
|
transformed_cache/train_tokenized_00002_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26f696795dee6c02fe594b04ac52f23c4d3b0d4f7daa66f42b9798a2abb1cf93
|
3 |
+
size 7876672
|
transformed_cache/train_tokenized_00003_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d477a0a9b8ff7cf1da7281b895b11e05affe217d81f44fcfc68eab70ecf0b3e
|
3 |
+
size 8214000
|
transformed_cache/train_tokenized_00004_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0b30406de2a4fb0ae56a08aca3a7457d6d63aa0ac66a7b950c84cd4252de96ed
|
3 |
+
size 8318008
|
transformed_cache/train_tokenized_00005_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:19c3e9c2abf180bda95e87458b6f610b45173b3eb4d7c10e37db10ee1b07cf64
|
3 |
+
size 8823040
|
transformed_cache/train_tokenized_00006_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a714384172fed68834fc866a5a799ffe9cb301598b41d5433d24a74727fa15a4
|
3 |
+
size 7516936
|
transformed_cache/train_tokenized_00007_of_00008.arrow
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a15c328603fef8a816eae22386f22485b37040adc7237397620e129656c9a565
|
3 |
+
size 7529536
|