Microsoft Windows [版本 10.0.19045.4170] (c) Microsoft Corporation。保留所有权利。 C:\Users\Lenovo>cd C:\Users\Lenovo\Desktop\wxy\CPT-master\finetune\generation C:\Users\Lenovo\Desktop\wxy\CPT-master\finetune\generation>python run_gen.py --model_path C:\Users\Lenovo\.cache\huggingface\hub\models--fnlp--cpt-large\snapshots\f07323ad5818364d47fc17cc4088072cd2f5f46d --dataset adgen --data_dir demo_data train validation test 03/22/2024 09:51:20 - WARNING - __main__ - Process rank: 0, device: cuda:0, n_gpu: 1distributed training: True, 16-bits training: False 03/22/2024 09:51:20 - INFO - __main__ - Training/evaluation parameters Seq2SeqTrainingArguments( _n_gpu=1, accelerator_config={'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True}, adafactor=False, adam_beta1=0.9, adam_beta2=0.999, adam_epsilon=1e-08, auto_find_batch_size=False, bf16=False, bf16_full_eval=False, data_seed=None, dataloader_drop_last=False, dataloader_num_workers=0, dataloader_persistent_workers=False, dataloader_pin_memory=True, dataloader_prefetch_factor=None, ddp_backend=None, ddp_broadcast_buffers=None, ddp_bucket_cap_mb=None, ddp_find_unused_parameters=None, ddp_timeout=1800, debug=[], deepspeed=None, disable_tqdm=False, dispatch_batches=None, do_eval=True, do_predict=True, do_train=True, eval_accumulation_steps=None, eval_delay=0, eval_steps=None, evaluation_strategy=epoch, fp16=False, fp16_backend=auto, fp16_full_eval=False, fp16_opt_level=O1, fsdp=[], fsdp_config={'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False}, fsdp_min_num_params=0, fsdp_transformer_layer_cls_to_wrap=None, full_determinism=False, generation_config=None, generation_max_length=None, generation_num_beams=None, gradient_accumulation_steps=1, gradient_checkpointing=False, gradient_checkpointing_kwargs=None, greater_is_better=None, group_by_length=False, half_precision_backend=auto, hub_always_push=False, hub_model_id=None, hub_private_repo=False, hub_strategy=every_save, hub_token=, ignore_data_skip=False, include_inputs_for_metrics=False, include_num_input_tokens_seen=False, include_tokens_per_second=False, jit_mode_eval=False, label_names=None, label_smoothing_factor=0.0, learning_rate=2e-05, length_column_name=length, load_best_model_at_end=False, local_rank=0, log_level=passive, log_level_replica=warning, log_on_each_node=True, logging_dir=output/adgen/6\runs\Mar22_09-51-19_DESKTOP-PC6Q6P1, logging_first_step=False, logging_nan_inf_filter=True, logging_steps=500, logging_strategy=steps, lr_scheduler_kwargs={}, lr_scheduler_type=linear, max_grad_norm=1.0, max_steps=-1, metric_for_best_model=None, mp_parameters=, neftune_noise_alpha=None, no_cuda=False, num_train_epochs=10.0, optim=adamw_torch, optim_args=None, output_dir=output/adgen/6, overwrite_output_dir=True, past_index=-1, per_device_eval_batch_size=6, per_device_train_batch_size=6, predict_with_generate=True, prediction_loss_only=False, push_to_hub=False, push_to_hub_model_id=None, push_to_hub_organization=None, push_to_hub_token=, ray_scope=last, remove_unused_columns=True, report_to=[], resume_from_checkpoint=None, run_name=output/adgen/6, save_on_each_node=False, save_only_model=False, save_safetensors=True, save_steps=500, save_strategy=no, save_total_limit=None, seed=6000, skip_memory_metrics=True, sortish_sampler=False, split_batches=None, tf32=None, torch_compile=False, torch_compile_backend=None, torch_compile_mode=None, torchdynamo=None, tpu_metrics_debug=False, tpu_num_cores=None, use_cpu=False, use_ipex=False, use_legacy_prediction_loop=False, use_mps_device=False, warmup_ratio=0.0, warmup_steps=0, weight_decay=0.0, ) loading file vocab.txt loading file added_tokens.json loading file special_tokens_map.json loading file tokenizer_config.json loading file tokenizer.json loading configuration file C:\Users\Lenovo\.cache\huggingface\hub\models--fnlp--cpt-large\snapshots\f07323ad5818364d47fc17cc4088072cd2f5f46d\config.json Model config BartConfig { "_name_or_path": "C:\\Users\\Lenovo\\.cache\\huggingface\\hub\\models--fnlp--cpt-large\\snapshots\\f07323ad5818364d47fc17cc4088072cd2f5f46d", "activation_dropout": 0.1, "activation_function": "gelu", "add_bias_logits": false, "add_final_layer_norm": false, "architectures": [ "BartForConditionalGeneration" ], "attention_dropout": 0.1, "bos_token_id": 101, "classif_dropout": 0.1, "classifier_dropout": 0.0, "d_model": 1024, "decoder_attention_heads": 16, "decoder_ffn_dim": 4096, "decoder_layerdrop": 0.0, "decoder_layers": 4, "decoder_start_token_id": 102, "dropout": 0.1, "early_stopping": true, "encoder_attention_heads": 16, "encoder_ffn_dim": 4096, "encoder_layerdrop": 0.0, "encoder_layers": 24, "eos_token_id": 102, "forced_eos_token_id": 102, "gradient_checkpointing": false, "id2label": { "0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2" }, "init_std": 0.02, "is_encoder_decoder": true, "label2id": { "LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2 }, "max_position_embeddings": 1024, "model_type": "bart", "no_repeat_ngram_size": 3, "normalize_before": false, "normalize_embedding": true, "num_beams": 4, "num_hidden_layers": 24, "pad_token_id": 0, "scale_embedding": false, "task_specific_params": { "summarization": { "length_penalty": 1.0, "max_length": 128, "min_length": 12, "num_beams": 4 }, "summarization_cnn": { "length_penalty": 2.0, "max_length": 142, "min_length": 56, "num_beams": 4 }, "summarization_xsum": { "length_penalty": 1.0, "max_length": 62, "min_length": 11, "num_beams": 6 } }, "tokenizer_class": "BertTokenizer", "transformers_version": "4.38.1", "use_cache": true, "vocab_size": 51271 } loading configuration file C:\Users\Lenovo\.cache\huggingface\hub\models--fnlp--cpt-large\snapshots\f07323ad5818364d47fc17cc4088072cd2f5f46d\config.json Model config BartConfig { "activation_dropout": 0.1, "activation_function": "gelu", "add_bias_logits": false, "add_final_layer_norm": false, "architectures": [ "BartForConditionalGeneration" ], "attention_dropout": 0.1, "bos_token_id": 101, "classif_dropout": 0.1, "classifier_dropout": 0.0, "d_model": 1024, "decoder_attention_heads": 16, "decoder_ffn_dim": 4096, "decoder_layerdrop": 0.0, "decoder_layers": 4, "decoder_start_token_id": 102, "dropout": 0.1, "early_stopping": true, "encoder_attention_heads": 16, "encoder_ffn_dim": 4096, "encoder_layerdrop": 0.0, "encoder_layers": 24, "eos_token_id": 102, "forced_eos_token_id": 102, "gradient_checkpointing": false, "id2label": { "0": "LABEL_0", "1": "LABEL_1", "2": "LABEL_2" }, "init_std": 0.02, "is_encoder_decoder": true, "label2id": { "LABEL_0": 0, "LABEL_1": 1, "LABEL_2": 2 }, "max_position_embeddings": 1024, "model_type": "bart", "no_repeat_ngram_size": 3, "normalize_before": false, "normalize_embedding": true, "num_beams": 4, "num_hidden_layers": 24, "pad_token_id": 0, "scale_embedding": false, "task_specific_params": { "summarization": { "length_penalty": 1.0, "max_length": 128, "min_length": 12, "num_beams": 4 }, "summarization_cnn": { "length_penalty": 2.0, "max_length": 142, "min_length": 56, "num_beams": 4 }, "summarization_xsum": { "length_penalty": 1.0, "max_length": 62, "min_length": 11, "num_beams": 6 } }, "tokenizer_class": "BertTokenizer", "transformers_version": "4.38.1", "use_cache": true, "vocab_size": 51271 } loading weights file C:\Users\Lenovo\.cache\huggingface\hub\models--fnlp--cpt-large\snapshots\f07323ad5818364d47fc17cc4088072cd2f5f46d\model.safetensors Generate config GenerationConfig { "bos_token_id": 101, "decoder_start_token_id": 102, "early_stopping": true, "eos_token_id": 102, "forced_eos_token_id": 102, "no_repeat_ngram_size": 3, "num_beams": 4, "pad_token_id": 0 } All model checkpoint weights were used when initializing CPTForConditionalGeneration. All the weights of CPTForConditionalGeneration were initialized from the model checkpoint at C:\Users\Lenovo\.cache\huggingface\hub\models--fnlp--cpt-large\snapshots\f07323ad5818364d47fc17cc4088072cd2f5f46d. If your task is similar to the task the model of the checkpoint was trained on, you can already use CPTForConditionalGeneration for predictions without further training. Generation config file not found, using a generation config created from the model config. Map: 0%| | 0/3290 [00:00