Problem with CodeT5+

#1
by AngeloCurti22 - opened

When trying to do inference with 2b,6b and 16b models, the following warning comes:

CodeT5pForCausalLM has generative capabilities, as prepare_inputs_for_generation is explicitly overwritten. However, it doesn't directly inherit from GenerationMixin. From ๐Ÿ‘‰v4.50๐Ÿ‘ˆ onwards, PreTrainedModel will NOT inherit from GenerationMixin, and this model will lose the ability to call generate and other related functions.

  • If you're using trust_remote_code=True, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes
  • If you are the owner of the model architecture code, please modify your model class such that it inherits from GenerationMixin (after PreTrainedModel, otherwise you'll get an exception).
  • If you are not the owner of the model architecture class, please contact the model code owner to update it.

Followed by the error:
AssertionError: Config has to be initialized with encoder and decoder config
Hope someone can help!

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment