hf_synth_trans / myerrors_1552.out
aatherton2024's picture
pushing major changes
8a736a7
raw
history blame
5.28 kB
Map: 0%| | 0/7292 [00:00<?, ? examples/s] Map: 27%|β–ˆβ–ˆβ–‹ | 2000/7292 [00:00<00:00, 9056.99 examples/s] Map: 82%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 6000/7292 [00:00<00:00, 17405.96 examples/s] Map: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 7292/7292 [00:00<00:00, 17292.17 examples/s]
Map: 0%| | 0/1001 [00:00<?, ? examples/s] Map: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1001/1001 [00:00<00:00, 28511.75 examples/s]
Map: 0%| | 0/1001 [00:00<?, ? examples/s] Map: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 1001/1001 [00:00<00:00, 30926.57 examples/s]
Downloading (…)lve/main/config.json: 0%| | 0.00/570 [00:00<?, ?B/s] Downloading (…)lve/main/config.json: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 570/570 [00:00<00:00, 1.74MB/s]
Downloading model.safetensors: 0%| | 0.00/436M [00:00<?, ?B/s] Downloading model.safetensors: 2%|▏ | 10.5M/436M [00:00<00:05, 71.5MB/s] Downloading model.safetensors: 5%|▍ | 21.0M/436M [00:00<00:04, 86.4MB/s] Downloading model.safetensors: 10%|β–‰ | 41.9M/436M [00:00<00:03, 102MB/s] Downloading model.safetensors: 14%|β–ˆβ– | 62.9M/436M [00:00<00:03, 108MB/s] Downloading model.safetensors: 19%|β–ˆβ–‰ | 83.9M/436M [00:00<00:03, 111MB/s] Downloading model.safetensors: 24%|β–ˆβ–ˆβ– | 105M/436M [00:00<00:02, 112MB/s] Downloading model.safetensors: 29%|β–ˆβ–ˆβ–‰ | 126M/436M [00:01<00:02, 113MB/s] Downloading model.safetensors: 34%|β–ˆβ–ˆβ–ˆβ–Ž | 147M/436M [00:01<00:02, 114MB/s] Downloading model.safetensors: 39%|β–ˆβ–ˆβ–ˆβ–Š | 168M/436M [00:01<00:02, 114MB/s] Downloading model.safetensors: 43%|β–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 189M/436M [00:01<00:02, 114MB/s] Downloading model.safetensors: 48%|β–ˆβ–ˆβ–ˆβ–ˆβ–Š | 210M/436M [00:01<00:02, 112MB/s] Downloading model.safetensors: 53%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 231M/436M [00:02<00:01, 114MB/s] Downloading model.safetensors: 58%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Š | 252M/436M [00:02<00:01, 113MB/s] Downloading model.safetensors: 63%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–Ž | 273M/436M [00:02<00:01, 114MB/s] Downloading model.safetensors: 67%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 294M/436M [00:02<00:01, 115MB/s] Downloading model.safetensors: 72%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 315M/436M [00:02<00:01, 114MB/s] Downloading model.safetensors: 77%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 336M/436M [00:03<00:00, 113MB/s] Downloading model.safetensors: 82%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ– | 357M/436M [00:03<00:00, 115MB/s] Downloading model.safetensors: 87%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹ | 377M/436M [00:03<00:00, 115MB/s] Downloading model.safetensors: 91%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–| 398M/436M [00:03<00:00, 114MB/s] Downloading model.safetensors: 96%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–‹| 419M/436M [00:03<00:00, 115MB/s] Downloading model.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 436M/436M [00:03<00:00, 115MB/s] Downloading model.safetensors: 100%|β–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆβ–ˆ| 436M/436M [00:03<00:00, 112MB/s]
You're using a GPT2TokenizerFast tokenizer. Please note that with a fast tokenizer, using the `__call__` method is faster than using a method to encode the text followed by a call to the `pad` method to get a padded encoding.
Traceback (most recent call last):
File "/mnt/storage/aatherton/hf_synth_trans/synth_translation.py", line 131, in <module>
trainer.evaluate(max_length=max_length)
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/transformers/trainer_seq2seq.py", line 159, in evaluate
return super().evaluate(eval_dataset, ignore_keys=ignore_keys, metric_key_prefix=metric_key_prefix)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/transformers/trainer.py", line 2972, in evaluate
output = eval_loop(
^^^^^^^^^^
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/transformers/trainer.py", line 3161, in evaluation_loop
loss, logits, labels = self.prediction_step(model, inputs, prediction_loss_only, ignore_keys=ignore_keys)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/transformers/trainer_seq2seq.py", line 282, in prediction_step
generated_tokens = self.model.generate(**inputs, **gen_kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/torch/utils/_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/transformers/generation/utils.py", line 1402, in generate
self._validate_model_class()
File "/mnt/storage/aatherton/anaconda3/envs/nmt/lib/python3.11/site-packages/transformers/generation/utils.py", line 1197, in _validate_model_class
raise TypeError(exception_message)
TypeError: The current model class (BertModel) is not compatible with `.generate()`, as it doesn't have a language model head. Please use one of the following classes instead: {'BertLMHeadModel'}