Finetuning model performance decreases when using memory_efficient_attenttion
#11 opened 30 days ago
by
hrushikesh1
Any Plan to release tensorflow based model?
1
#10 opened 4 months ago
by
undefined-x
adding to transformers officially?
#9 opened 4 months ago
by
pszemraj

Is flash-attention-2 suppported
1
#8 opened 6 months ago
by
Jack7777777
Xformer support for Qwen1.5B
3
#6 opened 7 months ago
by
le723z
backbone模型不开源吗?
10
#4 opened 10 months ago
by
JaheimLee
Disable trust_remote_code
14
#2 opened 11 months ago
by
veeravignesh