Minor fix for correct input forwarding
Browse files
model.py
CHANGED
@@ -323,13 +323,13 @@ class NeoBERTLMHead(NeoBERTPreTrainedModel):
|
|
323 |
):
|
324 |
|
325 |
output = self.model.forward(
|
326 |
-
input_ids,
|
327 |
-
position_ids,
|
328 |
-
max_seqlen,
|
329 |
-
cu_seqlens,
|
330 |
-
attention_mask,
|
331 |
-
output_hidden_states,
|
332 |
-
output_attentions,
|
333 |
)
|
334 |
logits = self.decoder(output.last_hidden_state)
|
335 |
|
@@ -380,13 +380,13 @@ class NeoBERTForSequenceClassification(NeoBERTPreTrainedModel):
|
|
380 |
):
|
381 |
|
382 |
output = self.model.forward(
|
383 |
-
input_ids,
|
384 |
-
position_ids,
|
385 |
-
max_seqlen,
|
386 |
-
cu_seqlens,
|
387 |
-
attention_mask,
|
388 |
-
output_hidden_states,
|
389 |
-
output_attentions,
|
390 |
)
|
391 |
hidden_states = output.last_hidden_state
|
392 |
|
|
|
323 |
):
|
324 |
|
325 |
output = self.model.forward(
|
326 |
+
input_ids=input_ids,
|
327 |
+
position_ids=position_ids,
|
328 |
+
max_seqlen=max_seqlen,
|
329 |
+
cu_seqlens=cu_seqlens,
|
330 |
+
attention_mask=attention_mask,
|
331 |
+
output_hidden_states=output_hidden_states,
|
332 |
+
output_attentions=output_attentions,
|
333 |
)
|
334 |
logits = self.decoder(output.last_hidden_state)
|
335 |
|
|
|
380 |
):
|
381 |
|
382 |
output = self.model.forward(
|
383 |
+
input_ids=input_ids,
|
384 |
+
position_ids=position_ids,
|
385 |
+
max_seqlen=max_seqlen,
|
386 |
+
cu_seqlens=cu_seqlens,
|
387 |
+
attention_mask=attention_mask,
|
388 |
+
output_hidden_states=output_hidden_states,
|
389 |
+
output_attentions=output_attentions,
|
390 |
)
|
391 |
hidden_states = output.last_hidden_state
|
392 |
|