Updating `hidden_states` value in line 869
Browse filesThis update is meant to solve the following error that occured when trying to finetune the model using peft :
RuntimeError Traceback (most recent call last)
in <cell line: 2>()
1 trainer.deprecated=True
----> 2 trainer.train()
21 frames
~/.cache/huggingface/modules/transformers_modules/inception-mbzuai/jais-13b-chat/96080d1c163804428c4792b8618c2d39661e9d7f/modeling_jais.py in forward(self, input_ids, past_key_values, attention_mask, token_type_ids, position_ids, head_mask, inputs_embeds, encoder_hidden_states, encoder_attention_mask, use_cache, output_attentions, output_hidden_states, return_dict)
867 else:
868 hidden_states = inputs_embeds
→ 869 hidden_states *= torch.tensor(
870 float(self.embeddings_scale), dtype=hidden_states.dtype, device=hidden_states.device
871 )
RuntimeError: a leaf Variable that requires grad is being used in an in-place operation.
#
- modeling_jais.py +5 -3
@@ -866,9 +866,11 @@ class JAISModel(JAISPreTrainedModel):
|
|
866 |
hidden_states = inputs_embeds + position_embeds
|
867 |
else:
|
868 |
hidden_states = inputs_embeds
|
869 |
-
hidden_states *= torch.tensor(
|
870 |
-
|
871 |
-
)
|
|
|
|
|
872 |
|
873 |
if token_type_ids is not None:
|
874 |
token_type_embeds = self.wte(token_type_ids)
|
|
|
866 |
hidden_states = inputs_embeds + position_embeds
|
867 |
else:
|
868 |
hidden_states = inputs_embeds
|
869 |
+
# hidden_states *= torch.tensor(
|
870 |
+
# float(self.embeddings_scale), dtype=hidden_states.dtype, device=hidden_states.device
|
871 |
+
# )
|
872 |
+
scale_factor_hidden = torch.tensor(float(self.embeddings_scale), dtype=hidden_states.dtype, device=hidden_states.device)
|
873 |
+
hidden_states = hidden_states * scale_factor_hidden
|
874 |
|
875 |
if token_type_ids is not None:
|
876 |
token_type_embeds = self.wte(token_type_ids)
|