Spaces:
Running
Running
Problem with CodeT5+
#1
by
AngeloCurti22
- opened
When trying to do inference with 2b,6b and 16b models, the following warning comes:
CodeT5pForCausalLM has generative capabilities, as prepare_inputs_for_generation
is explicitly overwritten. However, it doesn't directly inherit from GenerationMixin
. From ๐v4.50๐ onwards, PreTrainedModel
will NOT inherit from GenerationMixin
, and this model will lose the ability to call generate
and other related functions.
- If you're using
trust_remote_code=True
, you can get rid of this warning by loading the model with an auto class. See https://huggingface.co/docs/transformers/en/model_doc/auto#auto-classes - If you are the owner of the model architecture code, please modify your model class such that it inherits from
GenerationMixin
(afterPreTrainedModel
, otherwise you'll get an exception). - If you are not the owner of the model architecture class, please contact the model code owner to update it.
Followed by the error:
AssertionError: Config has to be initialized with encoder and decoder config
Hope someone can help!